Flower Research

MLSys ‘25

Photon: Federated LLM Pre-Training

Photon is the first full system for federated LLM training, enabling global-scale, low-bandwidth pre-training. It trains 7B models faster than baselines with better perplexity and 64x–512x less communication.

Links to Research Paper
arxivGoogle Scholar
Share

DEPT: Decoupled Embeddings for Pre-training Language Models

DEPT is a communication-efficient framework for federated LLM pre-training, enabling vocabulary-agnostic training across diverse data while cutting embedding memory 4–5x and improving perplexity up to 20%.

Links to Research Paper
arxivGoogle Scholar
Share

Supercharging Federated Learning with Flower and NVIDIA FLARE

Flower and FLARE integrate to boost the FL ecosystem—Flower supports FL development and research, while FLARE provides a production-ready runtime. Integration enables seamless, modification-free deployment.

Links to Research Paper
arxivGoogle Scholar
Share

Worldwide Federated Training of Language Models

WorldLM enables global federated LLM training across diverse legal and data settings via federations of federations, using partial model localization. It outperforms standard setups by up to 1.91×.

Links to Research Paper
arxivGoogle Scholar
Share

The Future of Large Language Model Pre-training is Federated

Photon enables scalable federated LLM training across institutions, unlocking underused global data and compute. It supports billion-scale models, matching centralized performance with robust convergence.

Links to Research Paper
arxivGoogle Scholar
Share

MedPerf: Open Benchmarking Platform for Medical Artificial Intelligence using Federated Evaluation

MedPerf is an open framework for benchmarking medical AI using federated evaluation, enabling privacy-first, large-scale, real-world testing of models across healthcare facilities.

Links to Research Paper
arxivGoogle Scholar
Share

Flower: A Friendly Federated Learning Framework

Flower is a new FL framework enabling large-scale, realistic federated learning on heterogeneous edge devices, supporting up to 15M clients and smooth migration from simulation to real devices.

Links to Research Paper
arxivGoogle Scholar
Share