Supervised Fine-Tuning (SFT) provides a baseline, but achieving alignment requires a more direct way to incorporate human judgment about response quality. This chapter shifts focus to building a Reward Model (RM), a critical component that learns to predict which AI-generated responses humans prefer.
You will learn the process of creating this RM, starting with the concept of learning directly from pairwise comparisons. We will cover methods for collecting human preference data, structuring these datasets, and selecting appropriate model architectures. The chapter details common training objectives, such as those based on the Bradley-Terry model, formulated as:
P(response1≻response2∣prompt)=σ(RM(prompt,response1)−RM(prompt,response2))
where σ is the sigmoid function and ≻ denotes preference.
Furthermore, we'll examine techniques for calibrating reward model scores to better reflect preference strength and discuss potential challenges encountered during reward modeling, including data quality issues and the risk of the model finding unintended shortcuts (reward hacking). By the end of this chapter, you will understand how to train a model that quantifies human preferences, setting the stage for reinforcement learning optimization.
3.1 Concept of Learning from Preferences
3.2 Human Preference Data Collection
3.3 Preference Dataset Formats and Structures
3.4 Reward Model Architectures
3.5 Training Objectives for Reward Models
3.6 Calibration of Reward Models
3.7 Potential Issues in Reward Modeling
3.8 Hands-on Practical: Training a Reward Model
© 2025 ApX Machine Learning