Model Monitoring: Detecting and Analyzing Data Issues

Note: Presentation times are in Coordinated Universal Time (UTC).

Website Maintenance Alert

Due to scheduled maintenance, the USENIX website may not be available on Monday, March 17, from 10:00 am–6:00 pm Pacific Daylight Time (UTC -7). We apologize for the inconvenience and thank you for your patience.

If you would like to register for NSDI '25, SREcon25 Americas, or PEPR '25, please complete your registration before or after this time period.

Wednesday, 13 October, 2021 - 16:3016:45

Dmitri Melikyan, Graphsignal, Inc.

Abstract: 

Machine learning models are deployed to production environments increasingly more often. Unlike traditional coded applications, serving ML models are fully data-driven. Unexpected input data, which the model has never seen in training and is not prepared to handle, may produce garbage model output that will be treated as valid by the rest of the application without producing any error or exception. For this reason production models introduce a new monitoring requirement, a model monitoring, that is not addressed by the existing tools. This talk focuses on monitoring and detection of production data issues, such as data anomalies and drift.

Dmitri Melikyan, Graphsignal, Inc.

Dmitri Melikyan is a founder and CEO at Graphsignal, an ML model monitoring company. He spent the last decade developing application monitoring and profiling tools used by thousands of developers and SREs.

SREcon21 Open Access Sponsored by Indeed

BibTeX

Presentation Video