Manifold: A Model-Agnostic Visual Debugging Tool for Machine Learning at Uber

Manifold: A Model-Agnostic Visual Debugging Tool for Machine Learning at Uber

  • January 23, 2019
Table of Contents

Manifold: A Model-Agnostic Visual Debugging Tool for Machine Learning at Uber

Uber built Manifold, a model-agnostic visualization tool for ML performance diagnosis and model debugging, to facilitate a more informed and actionable model iteration process. Machine learning (ML) is widely used across the Uber platform to support intelligent decision making and forecasting for features such as ETA prediction and fraud detection. For optimal results, we invest a lot of resources in developing accurate predictive ML models.

In fact, it’s typical for practitioners to devote 20 percent of their effort into building initial working models, and 80 percent of their effort improving model performance in what is known as the 20/80 split rule of ML model development. Traditionally, when data scientists develop models, they evaluate each model candidate using summary scores like log loss, area under curve (AUC), and mean absolute error (MAE). Although these metrics offer insights into how a model is performing, they do not convey much information regarding why a model is not performing well, and from there, how to improve its performance.

As such, model builders tend to rely on trial and error when determining how to improve their models.

Source: uber.com

Tags :
Share :
comments powered by Disqus

Related Posts

AI year in review

AI year in review

At Facebook, we think that artificial intelligence that learns in new, more efficient ways – much like humans do – can play an important role in bringing people together. That core belief helps drive our AI strategy, focusing our investments in long-term research related to systems that learn using real-world data, inspiring our engineers to share cutting-edge tools and platforms with the wider AI community, and ultimately demonstrating new ways to use the technology to benefit the world. In 2018, we made important progress in all these areas.

Read More