Amazon EC2 Accelerated Computing Instances, Amazon Web Services, 2024 (Amazon Web Services) - Official documentation detailing AWS's GPU instance offerings, including various GPU tiers, their specifications, networking options like EFA, and instance naming conventions.
NVIDIA H100 Tensor Core GPU Architecture, NVIDIA Corporation, 2022 (NVIDIA Corporation) - Technical whitepaper explaining the architecture of high-performance NVIDIA GPUs, covering features like HBM3, NVLink, and Tensor Cores, which are essential for large-scale AI training.
GPU platforms for Compute Engine, Google Cloud, 2024 (Google Cloud) - Official documentation outlining Google Cloud's GPU offerings for Compute Engine, including supported GPU models, their configurations, and how they relate to instance types like a2-highgpu.
Designing Machine Learning Systems, Chip Huyen, 2022 (O'Reilly Media) - A book providing system-level considerations for building ML products, with relevant chapters on selecting hardware, managing compute resources, and optimizing memory usage for deep learning workloads.