So far, we have explored methods like Monte Carlo and Temporal Difference learning that rely on maintaining a table mapping each state or state-action pair to a value. This approach works well for problems with a manageable number of states. However, many interesting problems involve state spaces that are enormous or even continuous, making tabular methods impractical due to memory and computational requirements.
This chapter addresses this challenge by introducing function approximation. Instead of storing exact values for every state, we will approximate the value function using a parameterized function with far fewer parameters than the number of states. We aim to generalize from seen states to unseen ones.
You will learn:
By the end of this chapter, you will understand how to apply RL algorithms to problems where the state space is too large to handle with simple tables.
6.1 Handling Large State Spaces
6.2 Value Function Approximation (VFA)
6.3 Feature Vectors for State Representation
6.4 Linear Methods for VFA
6.5 Gradient Descent for Parameter Learning
6.6 Semi-gradient TD Methods
6.7 Using Neural Networks for VFA
6.8 Practice: Applying Linear VFA
© 2025 ApX Machine Learning