A trained model artifact is not a production service. The process of deploying a model for inference introduces a distinct set of engineering challenges centered on performance and efficiency. While training prioritizes throughput over long jobs, inference services often must meet strict service-level objectives (SLOs) for single-request latency, such as a 99th percentile response time of p99<100ms. Meeting these goals requires specific optimizations at both the model and the infrastructure levels.
This chapter covers the methods for engineering and deploying high-performance inference systems. We will start by examining the architectural trade-offs between low latency and high throughput. You will then learn to apply post-training optimizations directly to your models, including:
With an optimized model, we will move to deployment using the NVIDIA Triton Inference Server. You will see how to manage multiple models, configure dynamic batching to improve GPU utilization, and implement safe rollout strategies like A/B testing and canary deployments for new model versions.
4.1 Architecting Inference Services for Latency and Throughput
4.2 Model Optimization with TensorRT and ONNX Runtime
4.3 Model Quantization Techniques: INT8 and FP8
4.4 Serving Multiple Models with NVIDIA Triton Inference Server
4.5 A/B Testing and Canary Deployments for Models
4.6 Hands-on Practical: Deploying an Optimized Model on Triton
© 2026 ApX Machine LearningEngineered with