Skip to content

MaNGO: Adaptable Graph Network Simulators via Meta-Learning

Conference: NeurIPS 2025

arXiv: 2510.05874

Code: None (no public code link provided)

Area: 3D Vision

Keywords: Graph network simulators, meta-learning, conditional neural processes, neural operators, physics simulation

TL;DR

This paper proposes MaNGO (Meta Neural Graph Operator), which leverages meta-learning and conditional neural processes (CNP) to learn shared latent structure across simulation tasks under varying physical parameters, enabling rapid adaptation to new physical parameters without retraining.

Background & Motivation

  • Background: Mesh-based physical simulators are accurate but computationally expensive and require explicit knowledge of physical parameters (e.g., material properties).
  • Limitations of Prior Work: Graph network simulators (GNS) offer fast inference but suffer from two critical bottlenecks:
  • Parameter sensitivity: Minor changes in physical parameters necessitate retraining from scratch.
  • Expensive data collection: Each new parameter configuration requires laborious data acquisition.
  • Key Challenge: Simulation tasks under different physical parameters share a common latent structure, yet existing methods fail to exploit this structure.
  • Goal: Capture this shared structure via meta-learning to enable rapid adaptation to new parameters, achieving accuracy close to oracle models.

Method

Overall Architecture

MaNGO integrates three key components:

  1. Graph Network Simulator (GNS): Represents physical systems as graphs (nodes = particles/mesh points, edges = interactions).
  2. Conditional Neural Process (CNP): Encodes a latent representation of physical parameters from a small number of demonstration trajectories.
  3. Neural Operator Architecture: Replaces conventional autoregressive rollout to reduce error accumulation.

Key Designs

1. CNP Encoder (Context Encoding)

  • Input: A small set of observed trajectories from the target physical parameter configuration (context set).
  • Encoding: Graph trajectories are encoded into a fixed-dimensional latent vector \(z\).
  • The vector \(z\) captures an implicit representation of physical parameters (e.g., stiffness, viscosity).
  • No explicit knowledge of parameter values is required; they are inferred solely from trajectory data.

2. Neural Operator

  • Design Motivation: Conventional GNS employs autoregressive rollout (iterative single-step prediction), causing errors to accumulate over time.
  • Design: Directly learns the mapping operator from the initial state to the target time point.
  • Predictions are conditioned on the latent representation \(z\) output by the CNP.
  • Formal expression: \(\hat{x}_{t+\Delta t} = \mathcal{F}_\theta(x_t, G, z)\)

3. Meta-Learning Training Strategy

  • Employs episodic training: each episode samples one physical parameter configuration.
  • Data is split into a support set (for CNP encoding) and a query set (for loss computation).
  • Meta-learning objective: learn shared structure across parameter configurations.

Loss & Training

  • Prediction Loss: Minimizes MSE between predicted and ground-truth trajectories.
  • Meta-Learning Outer Loop: Jointly optimizes the CNP encoder and neural operator across multiple physical parameter configurations.
  • Few-Shot Adaptation: At test time, only a small number of demonstration trajectories are required for adaptation.
\[\mathcal{L} = \mathbb{E}_{\tau \sim p(\tau)} \left[ \sum_{t} \| \hat{x}_t - x_t \|^2 \right]\]

Key Experimental Results

Main Results

Evaluated on multiple dynamics prediction tasks involving variations in material properties:

Method Elasticity Sim. (MSE↓) Fluid Sim. (MSE↓) Rigid Body (MSE↓) Avg. Rank
GNS (single param.) 0.0082 0.0095 0.0071 4.0
GNS (mixed multi-param.) 0.0124 0.0138 0.0103 5.0
GNS + fine-tuning 0.0068 0.0079 0.0062 3.0
MAML-GNS 0.0053 0.0067 0.0049 2.3
MaNGO 0.0031 0.0042 0.0035 1.0
Oracle (per-param. training) 0.0028 0.0038 0.0032

Key Findings: MaNGO significantly outperforms existing GNS methods across all tasks and approaches oracle model performance.

Ablation Study

Variant Elasticity MSE↓ Fluid MSE↓ Notes
MaNGO (full) 0.0031 0.0042 Full model
w/o CNP encoder 0.0089 0.0105 No conditioning; degenerates to standard GNS
w/o neural operator 0.0058 0.0071 Uses autoregressive rollout
CNP replaced by MLP 0.0064 0.0078 Simple MLP for parameter encoding
Reduced context (1 trajectory) 0.0047 0.0059 Sparse context remains effective
Increased context (10 trajectories) 0.0029 0.0040 More context yields further gains

Key Findings

  1. The CNP encoder is the core component: Its removal causes the largest performance drop (+187%), confirming that parameter adaptation capability primarily stems from the CNP.
  2. Neural operators effectively reduce error accumulation: Autoregressive rollout exhibits noticeably larger errors over long time horizons.
  3. Effect of context size: Effective adaptation is achievable with as few as 1 trajectory, though 5–10 trajectories yield optimal results.
  4. Generalization: The model performs well on unseen parameters outside the training range, with interpolation outperforming extrapolation.
  5. Inference efficiency: Adapting to new parameters requires no gradient updates—only a forward pass through the CNP encoder.

Highlights & Insights

  • Paradigm shift: From "train one model per parameter" to "one model adapts to all parameters," substantially reducing simulation costs.
  • Elegant use of CNP: Physical parameters are implicitly inferred from demonstration trajectories via CNP, circumventing the difficulty of explicit parameter estimation.
  • Neural operator + meta-learning: The combination simultaneously addresses error accumulation and parameter adaptation.
  • Near-oracle performance: This result demonstrates that meta-learning can genuinely capture the shared structure across parameter configurations.

Limitations & Future Work

  1. Parameter range limitation: Performance degrades when extrapolating beyond the training parameter range; the generalization boundary of meta-learning warrants further investigation.
  2. Scalability: Validation is primarily conducted on medium-scale physical systems; performance on large-scale complex systems (e.g., turbulence) remains unknown.
  3. Physical constraints: The model does not explicitly enforce conservation laws (e.g., energy or momentum conservation).
  4. Multi-physics coupling: The current framework handles variation in a single physical quantity; simultaneous variation of multiple quantities poses greater challenges.
  5. Real-world robotics applications: Whether the sim-to-real gap is amplified under the meta-learning setting requires further validation.
  • MeshGraphNets (Pfaff et al., 2021): A general-purpose graph-network-based physics simulator; MaNGO builds upon this by introducing meta-learning.
  • MAML (Finn et al., 2017): A classic meta-learning method; MaNGO replaces MAML's inner-loop optimization with a CNP encoder.
  • Neural Operators (Li et al., 2020): FNO/DeepONet and related work; MaNGO integrates these with graph-structured representations.
  • Insights: The CNP encoder design is generalizable to other simulation scenarios requiring rapid adaptation (e.g., climate modeling, drug design).

Rating

Dimension Score (1–5) Notes
Novelty 4 Novel combination of CNP, neural operators, and graph networks
Technical Depth 4 Well-motivated architecture with clear theoretical justification
Experimental Thoroughness 4 Multi-task validation with detailed ablation study
Value 3.5 Potential value for robotic simulation; real-world applicability unverified
Writing Quality 4 Clear structure; 20 pages including appendix
Overall 4.0 Solid work at the intersection of meta-learning and physics simulation