Measuring model drift is a critical component of an ML monitoring system. Drift is a change in distribution over time, measured for model inputs, outputs, and actuals of a model. Monitor for drift to identify if your models have grown stale, you have data quality issues, or if there are adversarial inputs in your model. Detecting drift with ML monitoring will help protect your models from performance degradation and allow you to better understand how to begin resolution.
Model drift is a key component of ML Monitoring. Dive into what drift is, why it’s important to keep track of, and how to troubleshoot and resolve the underlying issue when drift occurs.
Learn how to automate the life cycle of model construction, deployment, and ML monitoring by providing a set of novel high-level, declarative abstractions.
Explore machine learning specific risk factors with ML monitoring such as boundary erosion, entanglement, hidden feedback loops, undeclared consumers, data dependencies, and more.
An ML Test Score Rubric to quantify common real-world ML production issues.
An overview of the hardware and software infrastructure that support Facebook's machine learning initiatives at scale.
An outline of Uber's Machine Learning as a Service (MLaaS) as it operates globally and its scalability challenges.