Monitoring and improving the performance of machine learning models : how to use ModelDB and Spark to track and improve model performance over time / with Manasi Vartak & Jason Slepicka.

By: Contributor(s): Publisher: [Place of publication not identified] : O'Reilly, 2017Description: 1 online resource (1 streaming video file (35 min., 52 sec.)) : digital, sound, colorContent type:
  • two-dimensional moving image
Media type:
  • computer
  • video
Carrier type:
  • online resource
Other title:
  • How to use ModelDB and Spark to track and improve model performance over time
Subject(s): Genre/Form: LOC classification:
  • Q325.5
Online resources: Presenter, Manasi Vartak.Summary: "It's critical to have 'humans in the loop' when automating the deployment of machine learning (ML) models. Why? Because models often perform worse over time. This course covers the human directed safeguards that prevent poorly performing models from deploying into production and the techniques for evaluating models over time. We'll use ModelDB to capture the appropriate metrics that help you identify poorly performing models. We'll review the many factors that affect model performance (i.e., changing users and user preferences, stale data, etc.) and the variables that lose predictive power. We'll explain how to utilize classification and prediction scoring methods such as precision recall, ROC, and jaccard similarity. We'll also show you how ModelDB allows you to track provenance and metrics for model performance and health; how to integrate ModelDB with SparkML; and how to use the ModelDB APIs to store information when training models in Spark ML. Learners should have basic familiarity with the following: Scala or Python; Hadoop, Spark, or Pandas; SBT or Maven; cloud platforms like Amazon Web Services; Bash, Docker, and REST."--Resource description page.
Item type: eBooks
Star ratings
    Average rating: 0.0 (0 votes)
No physical items for this record

Title from title screen (Safari, viewed January 15, 2018).

Release date from resource description page (Safari, viewed January 15, 2018).

Presenter, Manasi Vartak.

"It's critical to have 'humans in the loop' when automating the deployment of machine learning (ML) models. Why? Because models often perform worse over time. This course covers the human directed safeguards that prevent poorly performing models from deploying into production and the techniques for evaluating models over time. We'll use ModelDB to capture the appropriate metrics that help you identify poorly performing models. We'll review the many factors that affect model performance (i.e., changing users and user preferences, stale data, etc.) and the variables that lose predictive power. We'll explain how to utilize classification and prediction scoring methods such as precision recall, ROC, and jaccard similarity. We'll also show you how ModelDB allows you to track provenance and metrics for model performance and health; how to integrate ModelDB with SparkML; and how to use the ModelDB APIs to store information when training models in Spark ML. Learners should have basic familiarity with the following: Scala or Python; Hadoop, Spark, or Pandas; SBT or Maven; cloud platforms like Amazon Web Services; Bash, Docker, and REST."--Resource description page.

Copyright © 2020 Alfaisal University Library. All Rights Reserved.
Tel: +966 11 2158948 Fax: +966 11 2157910 Email:
librarian@alfaisal.edu