Practical Deep Learning at Scale with MLflow

Practical Deep Learning at Scale with MLflow
-0 %
Der Artikel wird am Ende des Bestellprozesses zum Download zur Verfügung gestellt.
Bridge the gap between offline experimentation and online production
Sofort lieferbar | Lieferzeit: Sofort lieferbar

Unser bisheriger Preis:ORGPRICE: 34,97 €

Jetzt 33,58 €*

Artikel-Nr:
9781803242224
Veröffentl:
2022
Seiten:
288
Autor:
Yong Liu
eBook Typ:
EPUB
eBook Format:
Reflowable
Kopierschutz:
NO DRM
Sprache:
Englisch
Beschreibung:

Train, test, run, track, store, tune, deploy, and explain provenance-aware deep learning models and pipelines at scale with reproducibility using MLflowKey FeaturesFocus on deep learning models and MLflow to develop practical business AI solutions at scaleShip deep learning pipelines from experimentation to production with provenance trackingLearn to train, run, tune and deploy deep learning pipelines with explainability and reproducibilityBook DescriptionThe book starts with an overview of the deep learning (DL) life cycle and the emerging Machine Learning Ops (MLOps) field, providing a clear picture of the four pillars of deep learning: data, model, code, and explainability and the role of MLflow in these areas. From there onward, it guides you step by step in understanding the concept of MLflow experiments and usage patterns, using MLflow as a unified framework to track DL data, code and pipelines, models, parameters, and metrics at scale. You'll also tackle running DL pipelines in a distributed execution environment with reproducibility and provenance tracking, and tuning DL models through hyperparameter optimization (HPO) with Ray Tune, Optuna, and HyperBand. As you progress, you'll learn how to build a multi-step DL inference pipeline with preprocessing and postprocessing steps, deploy a DL inference pipeline for production using Ray Serve and AWS SageMaker, and finally create a DL explanation as a service (EaaS) using the popular Shapley Additive Explanations (SHAP) toolbox. By the end of this book, you'll have built the foundation and gained the hands-on experience you need to develop a DL pipeline solution from initial offline experimentation to final deployment and production, all within a reproducible and open source framework.What you will learnUnderstand MLOps and deep learning life cycle developmentTrack deep learning models, code, data, parameters, and metricsBuild, deploy, and run deep learning model pipelines anywhereRun hyperparameter optimization at scale to tune deep learning modelsBuild production-grade multi-step deep learning inference pipelinesImplement scalable deep learning explainability as a serviceDeploy deep learning batch and streaming inference servicesShip practical NLP solutions from experimentation to productionWho this book is forThis book is for machine learning practitioners including data scientists, data engineers, ML engineers, and scientists who want to build scalable full life cycle deep learning pipelines with reproducibility and provenance tracking using MLflow. A basic understanding of data science and machine learning is necessary to grasp the concepts presented in this book.

The book starts with an overview of the deep learning (DL) life cycle and the emerging Machine Learning Ops (MLOps) field, providing a clear picture of the four pillars of deep learning: data, model, code, and explainability and the role of MLflow in these areas.
From there onward, it guides you step by step in understanding the concept of MLflow experiments and usage patterns, using MLflow as a unified framework to track DL data, code and pipelines, models, parameters, and metrics at scale. You’ll also tackle running DL pipelines in a distributed execution environment with reproducibility and provenance tracking, and tuning DL models through hyperparameter optimization (HPO) with Ray Tune, Optuna, and HyperBand. As you progress, you’ll learn how to build a multi-step DL inference pipeline with preprocessing and postprocessing steps, deploy a DL inference pipeline for production using Ray Serve and AWS SageMaker, and finally create a DL explanation as a service (EaaS) using the popular Shapley Additive Explanations (SHAP) toolbox.
By the end of this book, you’ll have built the foundation and gained the hands-on experience you need to develop a DL pipeline solution from initial offline experimentation to final deployment and production, all within a reproducible and open source framework.

Kunden Rezensionen

Zu diesem Artikel ist noch keine Rezension vorhanden.
Helfen sie anderen Besuchern und verfassen Sie selbst eine Rezension.