In machine learning, data quality isn’t just a nice-to-have—it’s make or break. Bad data can silently derail your models, leading to poor predictions, wasted resources, and lost trust. In this talk, we’ll explore how to bring data validation into your ML pipelines using three powerful open-source tools: Great Expectations, Deequ, and TensorFlow Data Validation. We’ll look at how each tool helps catch issues like missing values, schema drift, and unexpected data distributions before they become bigger problems. You’ll see how they work, where they shine, and how to choose the right one for your workflow—whether you’re building batch pipelines, streaming systems, or end-to-end ML platforms. If you care about building reliable, production-ready ML systems, this session will give you the practical tools to keep your data in check.
Learn for free, join the best tech learning community for a price of a pumpkin latte.
Event notifications, weekly newsletter
Delayed access to all content
Immediate access to Keynotes & Panels
Access to Circle community platform
Immediate access to all content
Courses, quizes & certificates
Community chats