Having established a definition of Large Language Models, this chapter examines their internal workings at a high level. We will look at how these models process and generate text.
You will learn about:
This chapter provides a conceptual foundation for how LLMs operate without requiring deep mathematical or programming knowledge.
2.1 Representing Words: Tokens and Embeddings
2.2 Predicting the Next Word: The Core Idea
2.3 The Role of Training Data Size
2.4 Understanding Model Parameters
2.5 Introduction to Transformer Architecture (High-Level)
2.6 How Context Influences Generation
© 2025 ApX Machine Learning