Quiet Failures, Loud Consequences: Streaming ML Drift Detection in Practice

Breakout Session

A machine learning model in production is like a ship sailing blind, everything looks fine until it slams into a reef. And by then, it's too late.

This phenomenon, known as concept and model drift, is especially dangerous in real-time systems where decisions happen in milliseconds and rollback is usually not an option.

If not detected early, drift doesn’t just break your models — it misprices loans, misses fraud, and even risks lives.

This talk distills cutting-edge research and real production lessons into practical tools that can be apply today, even if the models are already in the wild. Based on ongoing PhD research and real-world implementations, we’ll walk through the following real live questions:

- How drift manifests in event-driven ML systems — and why traditional batch monitoring fails.

- Common algorithms for drift detection (i.e. DDM, EDDM, ADWIN, Page-Hinkley) — and how to benchmark them in streaming environments.

- An architecture for integrating drift-aware intelligence into Flink pipelines, with hooks for alerting, model retraining, or failover strategies.

- Lessons from production use cases, including trade-offs in detection latency, false positives, and system overhead.

Whether you're deploying ML models into dynamic data streams or just planning your streaming AI strategy, you'll leave with a blueprint for building drift-resilient ML pipelines — plus hands-on knowledge to detect, benchmark, and respond to drift before it becomes failure.


Dominique Ronde

Confluent