Training Robust Graph Neural Networks by Modeling Noise Dependencies¶
Conference: NeurIPS 2025 arXiv: 2502.19670 Code: GitHub Area: Optimization Keywords: GNN robustness, noise dependency, causal modeling, variational inference, data generating process
TL;DR¶
This paper proposes Dependency-Aware Graph Noise (DANG) and the DA-GNN framework, which model a causal dependency chain from node feature noise → graph structure noise → label noise, and employ variational inference to derive an ELBO objective for training GNNs robust to multi-source correlated noise.
Background & Motivation¶
One of the core challenges facing GNNs in real-world applications is data noise. Existing robust GNN methods address feature noise (AirGNN), structural noise (RSGNN, STABLE), or label noise (NRGNN, RTGNN) in isolation, all under the assumption that noise sources are mutually independent — the so-called Independent Feature Noise (IFN) assumption.
Core limitation: The IFN assumption is severely at odds with reality. In social networks, Bob creates a fake profile (feature noise) → Alice and Tom connect with Bob based on the fake profile (structural noise) → Alice and Tom's community labels are consequently corrupted (label noise). Such cascading noise dependencies are pervasive in social, e-commerce, and biological graph applications.
Fundamental limitations of prior work: Every existing robust GNN method assumes that at least one data source (features/structure/labels) is clean. For instance, AirGNN assumes a noise-free structure; RTGNN relies on structural information to mitigate label noise (even though the structure itself may be noisy); SG-GSR assumes clean labels. When all data sources are noisy, these assumptions are violated and performance degrades substantially.
Core Idea: Formally define DANG and its Data Generating Process (DGP), introducing three latent variables (noise variable \(\epsilon\), latent clean graph structure \(Z_A\), latent clean labels \(Z_Y\)) and establishing their causal relationships with observed variables (\(X\), \(A\), \(Y\)). A deep generative model, DA-GNN, is then designed to directly capture these causal relationships.
Method¶
Overall Architecture¶
DA-GNN is a variational inference-based encoder–decoder framework: - Inference encoders (\(\phi_1, \phi_2, \phi_3\)): infer latent variables \(Z_A, \epsilon, Z_Y\) from noisy observations - Generative decoders (\(\theta_1, \theta_2, \theta_3\)): model the process of generating observed data from latent variables - Trained by minimizing the negative ELBO
Key Designs¶
-
DGP for DANG: Six causal relationships are defined —
- \(X \leftarrow (\epsilon, Z_Y)\): the noise variable and true labels jointly produce the (potentially noisy) features
- \(A \leftarrow (Z_A, X)\): the latent clean structure and features jointly produce the (potentially noisy) edges
- \(A \leftarrow \epsilon\): the noise variable can also directly induce structural noise
- \(Y \leftarrow (Z_Y, X, A)\): true labels, features, and structure jointly produce the (potentially noisy) observed labels
A key property of this DGP is that no data source in the graph is assumed to be entirely clean.
-
Inference encoder design:
-
\(q_{\phi_1}(Z_A|X,A)\) (inferring clean graph structure): A GCN encoder computes node representations \(\mathbf{Z} = \text{GCN}_{\phi_1}(\mathbf{X}, \mathbf{A})\), and a latent graph is obtained via cosine similarity: \(\hat{p}_{ij} = \rho(s(\mathbf{Z}_i, \mathbf{Z}_j))\). A \(\gamma\)-hop subgraph similarity prior is used for regularization (corresponding to the KL divergence term). To avoid \(O(N^2)\) full-graph computation, a pre-defined proxy graph restricts the computational scope.
-
\(q_{\phi_3}(Z_Y|X,A)\) (inferring clean labels): A GCN classifier operating on the inferred clean graph \(\hat{\mathbf{A}}\). Homophily regularization encourages connected nodes to have similar predictions:
\(\mathcal{L}_{\text{hom}} = \sum_{i \in \mathcal{V}} \frac{\sum_{j \in \mathcal{N}_i} \hat{p}_{ij} \cdot kl(\hat{\mathbf{Y}}_j || \hat{\mathbf{Y}}_i)}{\sum_{j \in \mathcal{N}_i} \hat{p}_{ij}}\)
- \(q_{\phi_2}(\epsilon|X,A,Z_Y)\) (inferring the noise variable): Decomposed into structural noise \(\epsilon_A\) (edge cleanliness estimated via small-loss criterion during an early-learning phase) and feature noise \(\epsilon_X\) (inferred from \(X\) and \(Z_Y\) via an MLP, regularized toward a standard normal distribution). An EMA trick mitigates uncertainty in point estimates: \(\hat{p}_{ij}^{el} \leftarrow \xi \hat{p}_{ij}^{el} + (1-\xi)\hat{p}_{ij}^c\), with \(\xi=0.9\).
-
-
Generative decoder design:
- \(p_{\theta_1}(A|X,\epsilon,Z_A)\) (reconstructing noisy edges): An edge reconstruction loss that regularizes both predictions and labels to handle noisy supervision. Prediction regularization: \(\hat{p}_{ij}^{reg} = \theta_1 \hat{p}_{ij} + (1-\theta_1)s(\mathbf{X}_i, \mathbf{X}_j)\), which penalizes edge predictions when feature similarity is high (as spurious connections may arise from feature noise). Labels are smoothed to the interval \([0.9, 1]\).
- \(p_{\theta_2}(X|\epsilon,Z_Y)\) (reconstructing noisy features): An MLP decoder taking \(\epsilon_X\) and \(Z_Y\) as input, with samples drawn via the reparameterization trick.
- \(p_{\theta_3}(Y|X,A,Z_Y)\) (reconstructing noisy labels): A GCN classifier modeling the transition from clean to noisy labels.
Loss & Training¶
The total loss is a weighted sum of ELBO terms:
where \(\lambda_3=0.001\) is fixed (these three terms have limited impact on performance), \(\lambda_1\) and \(\lambda_2\) require tuning, and \(k\) (the proxy graph parameter) is the most critical hyperparameter.
Key Experimental Results¶
Main Results (Node classification accuracy % under DANG)¶
| Dataset | Noise Level | WSGNN | GraphGLOW | AirGNN | STABLE | RTGNN | SG-GSR | DA-GNN |
|---|---|---|---|---|---|---|---|---|
| Cora | Clean | 86.2 | 85.2 | 85.0 | 86.1 | 86.1 | 85.7 | 86.2 |
| Cora | DANG-10% | 80.7 | 79.7 | 79.7 | 82.2 | 81.8 | 82.7 | 82.9 |
| Cora | DANG-30% | 70.0 | 71.6 | 71.5 | 74.3 | 72.6 | 76.1 | 78.2 |
| Cora | DANG-50% | 55.9 | 59.6 | 56.2 | 62.8 | 60.9 | 64.3 | 69.7 |
| Photo | DANG-50% | 31.9 | 85.4 | 57.8 | 80.2 | 79.2 | 84.1 | 87.6 |
| Comp | DANG-50% | 39.6 | 80.1 | 44.1 | 68.8 | 69.4 | 78.6 | 82.2 |
Ablation Study (Incrementally removing causal relationships from the DGP)¶
| Configuration | Cora Clean | Cora DANG-30% | Cora DANG-50% | Description |
|---|---|---|---|---|
| Case 1 (IFN, remove all dependencies) | 84.6 | 68.3 | 55.2 | Degenerates to independent noise assumption |
| Case 2 (remove \(A \leftarrow X\)) | 84.8 | 68.5 | 56.1 | Ignores feature→structure noise propagation |
| Case 3 (remove only \(Y \leftarrow (X,A)\)) | 86.2 | 77.3 | 68.7 | Ignores label transition relationship |
| Proposed (full DANG) | 86.2 | 78.2 | 69.7 | Full causal modeling |
Real-World DANG Datasets¶
| Task | Dataset | Metric | SG-GSR (runner-up) | DA-GNN |
|---|---|---|---|---|
| Node classification | Auto + DANG | Accuracy | 62.0±1.1 | 61.4±0.4 |
| Node classification | Garden + DANG | Accuracy | 80.2±0.4 | 80.2±0.8 |
| Link prediction | Auto + DANG | ROC-AUC | 65.6±7.4 | 73.6±0.6 |
| Link prediction | Garden + DANG | ROC-AUC | 86.0±7.2 | 92.4±0.4 |
Key Findings¶
- The advantage of DA-GNN grows with noise level: On Cora DANG-50%, DA-GNN (69.7%) outperforms the runner-up SG-GSR (64.3%) by 5.4 percentage points.
- Every causal relationship in the DGP contributes: Incrementally removing causal edges leads to monotonically decreasing performance (69.7→68.7→56.1→55.2).
- Broad applicability: DA-GNN achieves state-of-the-art or competitive performance across five settings: DANG, feature-only noise, structure-only noise, label-only noise, and extreme noise (all three simultaneously).
- On the large-scale Arxiv graph, most baselines fail due to OOM; DA-GNN achieves 44.0% accuracy under DANG-50%.
Highlights & Insights¶
- DANG fills an important gap: The IFN assumption in existing robust GNN research is a widely accepted but unrealistic simplification; DANG provides a noise model that more faithfully reflects reality.
- Elegance of causal modeling: The DGP, expressed as a directed graphical model, cleanly defines six causal relationships, yielding a principled and theoretically grounded ELBO derivation.
- Practical new datasets: The Auto and Garden datasets, constructed from Amazon review data to simulate DANG in e-commerce settings, provide useful benchmarks for future research.
- Well-designed ablation: By incrementally removing causal edges from the DGP, the contribution of each dependency is clearly demonstrated.
Limitations & Future Work¶
- DANG does not consider the reverse dependency \(X \leftarrow A\) (structural noise feeding back into node features), which is also natural in certain settings.
- Training complexity is relatively high (three latent variables must be inferred jointly), which may limit scalability to large graphs.
- Proxy graph precomputation incurs \(O(N^2)\) initialization overhead; although optimized, it remains a bottleneck.
- The hyperparameter \(k\) (proxy graph parameter) requires tuning; while the search space is small, this adds a practical barrier.
- The performance gains on real-world DANG datasets are smaller than on synthetic ones, suggesting that synthetic settings may overestimate the method's advantage.
Related Work & Insights¶
- The approach is theoretically related to instance-dependent label noise (IDN) generative methods, but extending to the graph domain introduces additional latent variables and more complex causal relationships.
- WSGNN and GraphGLOW also employ variational inference and graph structure inference, but assume a noise-free graph, limiting their applicability.
- Insight: In other graph learning tasks such as recommendation systems and knowledge graphs, inter-noise dependencies are equally prevalent; the DANG modeling paradigm can be generalized accordingly.
Rating¶
- Novelty: ⭐⭐⭐⭐⭐ The definition of DANG and the DGP-based causal modeling methodology make important conceptual contributions.
- Experimental Thoroughness: ⭐⭐⭐⭐⭐ Covers synthetic/real DANG, individual noise types, and extreme noise settings, with detailed ablations.
- Writing Quality: ⭐⭐⭐⭐ The causal modeling section is clear, though the model instantiation details are dense and require careful reading.
- Value: ⭐⭐⭐⭐⭐ Advances the robust GNN research paradigm — from "assume some data source is clean" to "all data sources may be noisy."