Understanding LLM Model Sizes and Hardware Requirements
Practical Quantization for Large Language Models
Planning and Optimizing AI Infrastructure
Deploying Quantized LLMs for Efficient Inference
Agentic LLM Systems and Memory-Augmented Architectures