2409.03231
State-space models are accurate and efficient neural operators for dynamical systems
Zheyuan Hu, Nazanin Ahmadi Daryakenari, Qianli Shen, Kenji Kawaguchi, George Em Karniadakis
correctmedium confidence
- Category
- Not specified
- Journal tier
- Strong Field
- Processed
- Sep 28, 2025, 12:56 AM
- arXiv Links
- Abstract ↗PDF ↗
Audit review
The paper specifies the SSM/Mamba recurrence, its ZOH discretization, and the Mamba block architecture, and asserts linear-time inference and encoded temporal causality, contrasting this with generic neural operators like DeepONet; it also empirically demonstrates linear scaling and OOD robustness (Ex+/Ex-) rather than providing formal proofs. The model supplies rigorous versions of these points: an inductive causality proof for the Mamba scan, a clear O(N) time/activation-memory analysis, and an explicit construction showing DeepONet need not be causal. Thus the paper’s claims are substantively correct but informal, while the model provides the missing formal details. See the SSM/Mamba formulation and block description, the linear-time claim, the causality contrast with neural operators, and Ex+/Ex- evidence in the paper.
Referee report (LaTeX)
\textbf{Recommendation:} minor revisions \textbf{Journal Tier:} strong field \textbf{Justification:} This work persuasively demonstrates that Mamba, an SSM-based architecture, is a strong neural operator for dynamical systems: it scales linearly, respects temporal causality, and shows OOD robustness across diverse tasks. The claims about causality and efficiency are correct; adding brief formal statements would strengthen the theoretical underpinnings. The empirical methodology is broad and careful, with extensive baselines.