Deploying machine learning models as microservices using Docker : a REST-based architecture for serving ML model outputs at scale / with Hollin Wilkins & Jason Slepicka.

By: Contributor(s): Publisher: [Place of publication not identified] : O'Reilly, 2017Description: 1 online resource (1 streaming video file (24 min., 30 sec.)) : digital, sound, colorContent type:
  • two-dimensional moving image
Media type:
  • computer
  • video
Carrier type:
  • online resource
Other title:
  • REST-based architecture for serving ML model outputs at scale
Subject(s): Genre/Form: LOC classification:
  • Q325.5
Online resources: Presenter, Hollin Wilkins.Summary: "Modern applications running in the cloud often rely on REST-based microservices architectures by using Docker containers. Docker enables your applications to communicate between one another and to compose and scale various components. Data scientists use these techniques to efficiently scale their machine learning models to production applications. This video teaches you how to deploy machine learning models behind a REST API, to serve low latency requests from applications, without using a Spark cluster. In the process, you'll learn how to export models trained in SparkML; how to work with Docker, a convenient way to build, deploy, and ship application code for microservices; and how a model scoring service should support single on-demand predictions and bulk predictions. Learners should have basic familiarity with the following: Scala or Python; Hadoop, Spark, or Pandas; SBT or Maven; cloud platforms like Amazon Web Services; Bash, Docker, and REST."--Resource description page.
Item type: eBooks
Star ratings
    Average rating: 0.0 (0 votes)
No physical items for this record

Title from title screen (Safari, viewed January 15, 2018).

Release date from resource description page (Safari, viewed January 15, 2018).

Presenter, Hollin Wilkins.

"Modern applications running in the cloud often rely on REST-based microservices architectures by using Docker containers. Docker enables your applications to communicate between one another and to compose and scale various components. Data scientists use these techniques to efficiently scale their machine learning models to production applications. This video teaches you how to deploy machine learning models behind a REST API, to serve low latency requests from applications, without using a Spark cluster. In the process, you'll learn how to export models trained in SparkML; how to work with Docker, a convenient way to build, deploy, and ship application code for microservices; and how a model scoring service should support single on-demand predictions and bulk predictions. Learners should have basic familiarity with the following: Scala or Python; Hadoop, Spark, or Pandas; SBT or Maven; cloud platforms like Amazon Web Services; Bash, Docker, and REST."--Resource description page.

Copyright © 2020 Alfaisal University Library. All Rights Reserved.
Tel: +966 11 2158948 Fax: +966 11 2157910 Email:
librarian@alfaisal.edu