When working with functions of a single variable, like f(x), we used notations like dxdf or f′(x) to represent the derivative. Now that we're dealing with functions of multiple variables, such as f(x,y) or even f(x1,x2,…,xn), we need a way to specify which variable we are differentiating with respect to, while holding the others constant.
To distinguish partial derivatives from the ordinary derivatives we saw earlier, we use a different symbol: ∂. This symbol is often called the "curly d," "del," or simply "partial." It signals that we are performing a partial differentiation.
The most common notation you'll encounter, similar to the dxdy notation for ordinary derivatives, uses the ∂ symbol.
If we have a function f(x,y), its partial derivative with respect to x is written as:
∂x∂fThis notation reads as "the partial derivative of f with respect to x." It explicitly tells us:
Similarly, the partial derivative of f with respect to y is written as:
∂y∂fIf the function output is assigned to another variable, say z=f(x,y), you might also see the notation written as ∂x∂z or ∂y∂z.
Another common and often more compact notation uses subscripts to indicate the variable of differentiation. For the same function f(x,y):
This notation is especially convenient when writing out the gradient vector, which we'll see shortly.
Let's consider a simple function g(a,b)=3a2b+5b3.
(We'll learn how to calculate these in the next section, but for now, focus on understanding what the notation represents).
Just like with ordinary derivatives, we often want to evaluate a partial derivative at a specific point. If we want to find the partial derivative of f(x,y) with respect to x at the point (x0,y0), we can write it as:
∂x∂f(x0,y0)or∂x∂f(x0,y0)orfx(x0,y0)All these notations mean the same thing: calculate the partial derivative expression with respect to x, and then substitute x=x0 and y=y0 into the result.
Understanding this notation is fundamental because it's the language used to describe how complex functions, like the cost functions in machine learning models, change as their individual inputs or parameters are adjusted.
© 2025 ApX Machine Learning