Academic

Two Teachers Better Than One: Hardware-Physics Co-Guided Distributed Scientific Machine Learning

arXiv:2603.09032v1 Announce Type: new Abstract: Scientific machine learning (SciML) is increasingly applied to in-field processing, controlling, and monitoring; however, wide-area sensing, real-time demands, and strict energy and reliability constraints make centralized SciML implementation impractical. Most SciML models assume raw data aggregation at a central node, incurring prohibitively high communication latency and energy costs; yet, distributing models developed for general-purpose ML often breaks essential physical principles, resulting in degraded performance. To address these challenges, we introduce EPIC, a hardware- and physics-co-guided distributed SciML framework, using full-waveform inversion (FWI) as a representative task. EPIC performs lightweight local encoding on end devices and physics-aware decoding at a central node. By transmitting compact latent features rather than high-volume raw data and by using cross-attention to capture inter-receiver wavefield coupling,

arXiv:2603.09032v1 Announce Type: new Abstract: Scientific machine learning (SciML) is increasingly applied to in-field processing, controlling, and monitoring; however, wide-area sensing, real-time demands, and strict energy and reliability constraints make centralized SciML implementation impractical. Most SciML models assume raw data aggregation at a central node, incurring prohibitively high communication latency and energy costs; yet, distributing models developed for general-purpose ML often breaks essential physical principles, resulting in degraded performance. To address these challenges, we introduce EPIC, a hardware- and physics-co-guided distributed SciML framework, using full-waveform inversion (FWI) as a representative task. EPIC performs lightweight local encoding on end devices and physics-aware decoding at a central node. By transmitting compact latent features rather than high-volume raw data and by using cross-attention to capture inter-receiver wavefield coupling, EPIC significantly reduces communication cost while preserving physical fidelity. Evaluated on a distributed testbed with five end devices and one central node, and across 10 datasets from OpenFWI, EPIC reduces latency by 8.9$\times$ and communication energy by 33.8$\times$, while even improving reconstruction fidelity on 8 out of 10 datasets.

Executive Summary

The article introduces EPIC, a novel distributed Scientific Machine Learning (SciML) framework that integrates hardware and physics constraints into the model design. By leveraging lightweight local encoding on edge devices and physics-aware decoding at a central node—transmitting compact latent features instead of raw data and utilizing cross-attention to capture wavefield coupling—EPIC achieves significant reductions in communication latency (8.9×) and energy (33.8×), while simultaneously improving reconstruction fidelity on the majority of tested datasets. This is a pivotal advancement in addressing the dual challenges of scalability and physical consistency in distributed SciML, particularly for resource-constrained environments. The evaluation on real-world OpenFWI datasets lends substantial credibility to the claims.

Key Points

  • EPIC reduces communication latency by 8.9× and energy by 33.8×

Merits

Innovative Integration

EPIC uniquely co-guides hardware and physics constraints into distributed learning architecture, avoiding the trade-off between distributed scalability and physical fidelity.

Empirical Validation

Strong empirical results across 10 datasets demonstrate tangible improvements in both efficiency and accuracy.

Demerits

Scope Limitation

Evaluation is confined to full-waveform inversion (FWI); applicability to other SciML domains (e.g., climate modeling, biomedical imaging) remains unproven.

Generalizability Concerns

The framework’s cross-attention mechanism may incur computational overhead in highly heterogeneous hardware environments.

Expert Commentary

EPIC represents a paradigm shift in distributed SciML by operationalizing the coexistence of computational constraints and physical laws. Historically, distributed ML architectures have been designed with a focus on data aggregation and general-purpose optimization, often at the expense of physical consistency. EPIC’s architecture—by treating encoding as a preprocessing step aligned with physical observables and decoding as a domain-aware reconstruction—offers a sophisticated solution to a persistent bottleneck. The cross-attention mechanism, while mathematically elegant, warrants further scrutiny for scalability under variable latency or bandwidth conditions. Moreover, the absence of comparative benchmarks against state-of-the-art distributed SciML variants (e.g., FedAvgML, ML-Edge) limits the ability to fully contextualize EPIC’s relative contribution. Nevertheless, the results are compelling and suggest a viable pathway toward sustainable, physics-aligned distributed learning. This work should be viewed as foundational for future research in energy-efficient scientific inference.

Recommendations

  • Future work should extend EPIC’s framework to additional SciML applications beyond FWI, such as seismic inversion, fluid dynamics, or quantum simulation.
  • Comparative evaluations against alternative distributed learning architectures should be conducted to quantify relative performance gains and identify potential interoperability challenges.

Sources