A trained machine learning model generates value only when applications can use it to make predictions. This chapter covers the technical processes required to move a model from a local environment into a production system where it can serve those predictions. We will examine the complete path from a saved model file to an operational, monitored service.
You will learn to:
The chapter concludes with a hands-on section where you will containerize a simple model and prepare it for deployment.
5.1 What is Model Deployment?
5.2 Containerization with Docker
5.3 Creating a Model API with Flask
5.4 Deployment Patterns: Online vs. Batch Prediction
5.5 Introduction to Model Registries
5.6 Fundamentals of Model Monitoring
5.7 Hands-on Practical: Package and Deploy a Model as a Docker Container