Academic

BlazeFL: Fast and Deterministic Federated Learning Simulation

arXiv:2604.03606v1 Announce Type: new Abstract: Federated learning (FL) research increasingly relies on single-node simulations with hundreds or thousands of virtual clients, making both efficiency and reproducibility essential. Yet parallel client training often introduces nondeterminism through shared random state and scheduling variability, forcing researchers to trade throughput for reproducibility or to implement custom control logic within complex frameworks. We present BlazeFL, a lightweight framework for single-node FL simulation that alleviates this trade-off through free-threaded shared-memory execution and deterministic randomness management. BlazeFL uses thread-based parallelism with in-memory parameter exchange between the server and clients, avoiding serialization and inter-process communication overhead. To support deterministic execution, BlazeFL assigns isolated random number generator (RNG) streams to clients. Under a fixed software/hardware stack, and when stochasti

K
Kitsuya Azuma, Takayuki Nishio
· · 1 min read · 18 views

arXiv:2604.03606v1 Announce Type: new Abstract: Federated learning (FL) research increasingly relies on single-node simulations with hundreds or thousands of virtual clients, making both efficiency and reproducibility essential. Yet parallel client training often introduces nondeterminism through shared random state and scheduling variability, forcing researchers to trade throughput for reproducibility or to implement custom control logic within complex frameworks. We present BlazeFL, a lightweight framework for single-node FL simulation that alleviates this trade-off through free-threaded shared-memory execution and deterministic randomness management. BlazeFL uses thread-based parallelism with in-memory parameter exchange between the server and clients, avoiding serialization and inter-process communication overhead. To support deterministic execution, BlazeFL assigns isolated random number generator (RNG) streams to clients. Under a fixed software/hardware stack, and when stochastic operators consume BlazeFL-managed generators, this design yields bitwise-identical results across repeated high-concurrency runs in both thread-based and process-based modes. In CIFAR-10 image-classification experiments, BlazeFL substantially reduces execution time relative to a widely used open-source baseline, achieving up to 3.1$\times$ speedup on communication-dominated workloads while preserving a lightweight dependency footprint. Our open-source implementation is available at: https://github.com/kitsuyaazuma/blazefl.

Executive Summary

BlazeFL is a lightweight framework for single-node federated learning simulation that addresses the trade-off between efficiency and reproducibility. By utilizing free-threaded shared-memory execution and deterministic randomness management, BlazeFL achieves significant speedup and bitwise-identical results across repeated high-concurrency runs. The framework's design enables substantial reductions in execution time relative to a widely used open-source baseline, while preserving a lightweight dependency footprint. BlazeFL's open-source implementation is now available, offering researchers a valuable tool for FL simulation and evaluation. The implications of this work are significant, as it enables more efficient and reproducible FL research, which is critical for the widespread adoption of FL in real-world applications.

Key Points

  • BlazeFL is a lightweight framework for single-node FL simulation
  • Free-threaded shared-memory execution and deterministic randomness management enable significant speedup and reproducibility
  • BlazeFL achieves up to 3.1$ imes$ speedup on communication-dominated workloads relative to a widely used open-source baseline

Merits

Strength in Design

BlazeFL's design addresses the trade-off between efficiency and reproducibility, enabling significant speedup and bitwise-identical results across repeated high-concurrency runs.

Efficient Execution

BlazeFL's free-threaded shared-memory execution and deterministic randomness management enable substantial reductions in execution time relative to a widely used open-source baseline.

Reproducibility

BlazeFL's design ensures bitwise-identical results across repeated high-concurrency runs, making it an ideal tool for FL research and evaluation.

Demerits

Limited Scalability

BlazeFL's design may not be easily scalable to larger numbers of clients or more complex FL scenarios, potentially limiting its applicability to real-world FL deployments.

Dependence on Software/Hardware Stack

BlazeFL's deterministic execution requires a fixed software/hardware stack, which may limit its portability and applicability to different environments.

Expert Commentary

BlazeFL is a significant contribution to the field of FL research, as it addresses the trade-off between efficiency and reproducibility. The framework's design and implementation are well-suited for evaluating FL models and algorithms, making it a valuable tool for researchers and practitioners. While BlazeFL may have limitations in terms of scalability and portability, its impact on the field of FL research is substantial. Furthermore, the development of BlazeFL highlights the need for more efficient and reproducible FL research, which may inform policy decisions regarding the adoption of FL in various industries and applications.

Recommendations

  • Researchers and practitioners should consider using BlazeFL for FL simulation and evaluation due to its efficient execution and reproducibility.
  • The development of BlazeFL highlights the need for more efficient and reproducible FL research, which may inform policy decisions regarding the adoption of FL in various industries and applications.

Sources

Original: arXiv - cs.LG