What is Federated Learning?
The Core Problem
Traditional ML requires centralizing all training data. For keyboard prediction, this means uploading every keystroke from millions of users. For healthcare models, hospitals must share patient records. Both face insurmountable barriers: users refuse to share typing patterns, and regulations prohibit transferring patient data. The data exists, but cannot be accessed conventionally.
How It Works
Instead of bringing data to the model, federated learning brings the model to the data. A central server sends the current model to participating devices (clients). Each client trains locally for several iterations, producing updated weights. Clients send only weight updates back. The server aggregates updates from all clients into an improved model by averaging weights. This cycle repeats until convergence. A single round might involve 10,000 devices training locally for 5 epochs, with aggregation every 10-30 minutes.
Why Simple Alternatives Fail
Why not anonymize data and centralize it? Anonymization is fragile: 87% of Americans can be uniquely identified from zip code, birth date, and gender. Even aggregate statistics leak information. Federated learning ensures raw data never leaves the device. The server sees only model updates, which are mathematically difficult to reverse-engineer into original training examples.