Stress-Test Node & Link Models with One Click: Meet HypNF

Written by hyperbole | Published 2025/09/10
Tech Story Tags: network-topology | hyp-nf | graph-benchmark | synthetic-graphs | hyperbolic-geometry | gnn-testing | homophily | graph-simulation

TLDRBy knob-twisting γ, β and ρ, the framework creates sparse or dense, assortative or disassortative nets - perfect for stress-testing new GNNs.via the TL;DR App

Table of Links

Abstract and 1. Introduction

  1. Related work

  2. HypNF Model

    3.1 HypNF Model

    3.2 The S1/H2 model

    3.3 Assigning labels to nodes

  3. HypNF benchmarking framework

  4. Experiments

    5.1 Parameter Space

    5.2 Machine learning models

  5. Results

  6. Conclusion, Acknowledgments and Disclosure of Funding, and References

A. Empirical validation of HypNF

B. Degree distribution and clustering control in HypNF

C. Hyperparameters of the machine learning models

D. Fluctuations in the performance of machine learning models

E. Homophily in the synthetic networks

F. Exploring the parameters’ space

3 HypNF Model

3.1 The S1/H2 model

3.2 The bipartite-S1/H2 model

3.3 Assigning labels to nodes

4 HypNF benchmarking framework

The HypNF benchmarking framework depicted in Fig. 1 combines the S1/H2 and bipartite-S1/H2 models within a unified similarity space. Additionally, it incorporates a method for label assignment. This integration facilitates the creation of networks exhibiting a wide range of structural properties and varying degrees of correlation between nodes and their features. Specifically, our framework allows us to control the following properties:

Leveraging the HypNF model with varying parameters, our benchmarking framework generates diverse graph-structured data. This allows for the evaluation of graph machine learning models on networks with different connectivity patterns and correlations between topology and node features. For tasks like NC and LP, the framework facilitates fair model comparisons, helping to assess a novel GNN against state-of-the-art architectures and providing insights into the data’s impact on performance.

Authors:

(1) Roya Aliakbarisani, this author contributed equally from Universitat de Barcelona & UBICS ([email protected]);

(2) Robert Jankowski, this author contributed equally from Universitat de Barcelona & UBICS ([email protected]);

(3) M. Ángeles Serrano, Universitat de Barcelona, UBICS & ICREA ([email protected]);

(4) Marián Boguñá, Universitat de Barcelona & UBICS ([email protected]).


This paper is available on arxiv under CC by 4.0 Deed (Attribution 4.0 International) license.


Written by hyperbole | Amplifying words and ideas to separate the ordinary from the extraordinary, making the mundane majestic.
Published by HackerNoon on 2025/09/10