π PhysicsEngine: Reduced-Order Neural Operators for Lagrangian Dynamics
By Hrishikesh Viswanath, Yue Chang, Julius Berner, Peter Yichen Chen, Aniket Bera
π Model Overview
GIOROM is a Reduced-Order Neural Operator Transformer designed for Lagrangian dynamics simulations on highly sparse graphs. The model enables hybrid Eulerian-Lagrangian learning by:
- Projecting Lagrangian inputs onto uniform grids with a Graph-Interaction-Operator.
- Predicting acceleration from sparse velocity inputs using past time windows with a Neural Operator Transformer.
- Learning physics from sparse inputs (n βͺ N) while allowing reconstruction at arbitrarily dense resolutions via an Integral Transform Model.
- Dataset Compatibility: This model is compatible with MPM-Verse-MaterialSim-Small/Elasticity3DSmall,
β Note: While the model can infer using an integral transform, this repository only provides weights for the time-stepper model that predicts acceleration.
π Available Model Variants
Each variant corresponds to a specific dataset, showcasing the reduction in particle count (n: reduced-order, N: full-order).
| Model Name | n (Reduced) | N (Full) | 
|---|---|---|
| giorom-3d-t-sand3d-long | 3.0K | 32K | 
| giorom-3d-t-water3d | 1.7K | 55K | 
| giorom-3d-t-elasticity | 2.6K | 78K | 
| giorom-3d-t-plasticine | 1.1K | 5K | 
| giorom-2d-t-water | 0.12K | 1K | 
| giorom-2d-t-sand | 0.3K | 2K | 
| giorom-2d-t-jelly | 0.2K | 1.9K | 
| giorom-2d-t-multimaterial | 0.25K | 2K | 
π‘ How It Works
πΉ Input Representation
The model predicts acceleration from past velocity inputs:
- Input Shape: - [n, D, W]- n: Number of particles (reduced-order, n βͺ N)
- D: Dimension (2D or 3D)
- W: Time window (past velocity states)
 
- Projected to a uniform latent space of size - [c^D, D]where:- c β {8, 16, 32}
- n - Ξ΄n β€ c^D β€ n + Ξ΄n
 
This allows the model to generalize physics across different resolutions and discretizations.
πΉ Prediction & Reconstruction
- The model learns physical dynamics on the sparse input representation.
- The integral transform model reconstructs dense outputs at arbitrary resolutions (not included in this repo).
- Enables highly efficient, scalable simulations without requiring full-resolution training.
π Usage Guide
1οΈβ£ Install Dependencies
pip install transformers huggingface_hub torch
git clone https://github.com/HrishikeshVish/GIOROM/
cd GIOROM
2οΈβ£ Load a Model
from models.giorom3d_T import PhysicsEngine
from models.config import TimeStepperConfig
time_stepper_config = TimeStepperConfig()
simulator = PhysicsEngine(time_stepper_config)
repo_id = "hrishivish23/giorom-3d-t-sand3d"
time_stepper_config = time_stepper_config.from_pretrained(repo_id)
simulator = simulator.from_pretrained(repo_id, config=time_stepper_config)
3οΈβ£ Run Inference
import torch
π Model Weights and Checkpoints
| Model Name | Model ID | 
|---|---|
| giorom-3d-t-sand3d-long | hrishivish23/giorom-3d-t-sand3d-long | 
| giorom-3d-t-water3d | hrishivish23/giorom-3d-t-water3d | 
π Training Details
π§ Hyperparameters
- Graph Interaction Operator layers: 4
- Transformer Heads: 4
- Embedding Dimension: 128
- Latent Grid Sizes: {8Γ8, 16Γ16, 32Γ32}
- Learning Rate: 1e-4
- Optimizer: Adamax
- Loss Function: MSE + Physics Regularization (Loss computed on Euler integrated outputs)
- Training Steps: 1M+ steps
π₯οΈ Hardware
- Trained on: NVIDIA RTX 3050
- Batch Size: 2
π Citation
If you use this model, please cite:
@article{viswanath2024reduced,
  title={Reduced-Order Neural Operators: Learning Lagrangian Dynamics on Highly Sparse Graphs},
  author={Viswanath, Hrishikesh and Chang, Yue and Berner, Julius and Chen, Peter Yichen and Bera, Aniket},
  journal={arXiv preprint arXiv:2407.03925},
  year={2024}
}
π¬ Contact
For questions or collaborations:
- π§βπ» Author: Hrishikesh Viswanath
- π§ Email: [email protected]
- π¬ Hugging Face Discussion: Model Page
π Related Work
- Neural Operators for PDEs: Fourier Neural Operators, Graph Neural Operators
- Lagrangian Methods: Material Point Methods, SPH, NCLAW, CROM, LiCROM
- Physics-Based ML: PINNs, GNS, MeshGraphNet
πΉ Summary
This model is ideal for fast and scalable physics simulations where full-resolution computation is infeasible. The reduced-order approach allows efficient learning on sparse inputs, with the ability to reconstruct dense outputs using an integral transform model (not included in this repo).
- Downloads last month
- 1
