Big Data SystemsLambda & Kappa ArchitecturesMedium⏱️ ~2 min

What is Kappa Architecture?

Kappa Architecture treats all data processing as streaming, eliminating the separate batch layer entirely. There is a single pipeline that reads from an immutable append only log like Kafka, computes stateful materialized views, and serves them with low latency. When you need to reprocess historical data to fix bugs or add new metrics, you simply replay the log through the same streaming logic, often at higher speed. This drastically simplifies operations because you write transformation logic once and maintain one code path. The key enablers are durable ordered logs with long retention periods, often weeks to months using tiered storage, and stream processors that support event time semantics, state snapshots, and recovery. LinkedIn handles multiple trillion messages per day through Kafka, with individual stream processors operating at millions of messages per second and subsecond to low second latencies for real time features like ranking signals and abuse detection. Reddit processes billions of events per day for experimentation and ads attribution, retaining streams long enough to allow replay for bug fixes and backfills. Kappa trades long term log retention costs and strong streaming correctness guarantees for architectural simplicity. You shift complexity into the streaming engine, which must handle exactly once processing, watermarking for late data, and robust state management. Reprocessing can spike compute and I/O costs significantly when replaying months of data, requiring 1.5 to 3 times capacity headroom. Choose Kappa when near real time is first class, you want to avoid duplicate logic, and replay based reprocessing is acceptable for your scale.
💡 Key Takeaways
Single streaming pipeline eliminates dual logic and reduces operational complexity, but shifts correctness burden to streaming engine for exactly once semantics and state management
Long term log retention of weeks to months enables replay based reprocessing, requiring tiered storage and capacity planning for 1.5 to 3 times headroom during backfills
LinkedIn processes multiple trillion messages per day with subsecond to low second latencies, using Kappa principles with durable event logs and stream processors for both real time and replay
Reprocessing months of data can take hours to days, potentially throttling live processing and causing downstream cascades if capacity is insufficient
Choose Kappa when near real time processing is first class, you want to write transformation logic once, and replay based reprocessing fits your operational model
📌 Examples
Reddit processes billions of events per day for experimentation and ads attribution with second level latency for real time detectors, retaining event streams long enough for replay during bug fixes and new metric backfills
Kappa reprocessing: spin up new consumer group reading from beginning timestamp, write to shadow materialized view, validate against canaries, atomically switch read traffic after validation
← Back to Lambda & Kappa Architectures Overview
What is Kappa Architecture? | Lambda & Kappa Architectures - System Overflow