Extrema
In this part of the course we work on the following skills:
- Locating and classifying the extrema of scalar fields.
- Applying Lagrange's multipliers method to optimize quantities with respect to constraints.
See also the exercises associated to this part of the course.
In the previous chapter we introduced various notions of differentials for higher dimensional functions (scalar fields, vector fields, paths, etc.). This part of the course is devoted to searching for extrema (minima / maxima) in various different scenarios. This extends what we already know for functions in
Extrema (minima / maxima / saddle)
Let
Definition
If
Definition
If
Collectively we call the these points the extrema of the scalar field. In the case of a scalar field defined on
To proceed it is convenient to connect the extrema with the behaviour of the gradient of the scalar field.
Theorem
If
Proof
Suppose
Observe that, here and in the subsequent text, we can always consider the case of
Definition (stationary point)
If
As we see in the inflection example, the converse of the above theorem fails in the sense that a stationary point might not be a minimum or a maximum. The motivates the following.
Definition (saddle point)
If
The quintessential saddle has the shape seen in the graph. However it might be similar to an inflection in 1D or more complicated using the possibilities available in higher dimension.
Hessian matrix
To proceed it is useful to develop the idea of a second order Taylor expansion in this higher dimensional setting. In particular this will allow us to identify the local behaviour close to stationary points. The main object for doing this is the Hessian matrix. Let
Observe that the Hessian matrix
for twice differentiable functions.
The Hessian matrix is defined analogously in any dimension.
Let
Observe that the Hessian matrix is a real symmetric matrix in any dimension. If
As an example, let
The point
Theorem
If
Proof
Multiplying the matrices we calculate that
as required.
Second order Taylor formula for scalar fields
First let's recall the first order Taylor approximation we saw before. If
If
Theorem (second order Taylor for scalar fields)
Let
in the sense that the error is
Proof
Let
Consequently
Since
Classifying stationary points
In order to classify the stationary points we will take advantage of the Hessian matrix and therefore we need to first understand the follow fact about real symmetric matrices.
Theorem
Let
for all all eigenvalues of are positive, for all all eigenvalues of are negative.
Proof
Since
In order to prove the other direction in the "if and only if" statement, observe that
Theorem (classifying stationary points)
Let
- All eigenvalues are positive
relative minimum at , - All eigenvalues are negative
relative maximum at , - Some positive, some negative
is a saddle point.
Proof
Let
Since
Attaining extreme values
Here we explore the extreme value theorem for continuous scalar fields. The argument will be in two parts: Firstly we show that continuity implies boundedness; Secondly we show that boundedness implies that the maximum and minimum are attained. We use the following notation for interval / rectangle / cuboid / tesseract, etc. If
We call this set a rectangle (independent of the dimension). As a first step it is convenient to know that all sequences in our setting have convergent subsequences.
Theorem
If
Proof
In order to prove the theorem we construct the subsequence. Firstly we divide
Theorem
Suppose that
Proof
Suppose the contrary: for all
We can now use the above result on the boundedness in order to show that the extreme values are actually obtained.
Theorem
Suppose that
Proof
By the boundedness theorem
Extrema with constraints (Lagrange multipliers)
We now consider a slightly different problem to the one earlier in this chapter. There we wished to find the extrema of a given scalar field. Here the general problem is to minimise or maximise a given scalar field
Suppose that a differentiable scalar field
In three dimensions a similar result holds. Suppose that a differentiable scalar field
and the
In higher dimensions and possibly with additional constraints we have the following general theorem.
Theorem (Lagrange multipliers)
Suppose that a differentiable scalar field
where
The Lagrange multiplier method is often stated and far less often proved.
Since the proof is rather involved we will follow this tradition here. See, for example, Chapter 14 of "A First Course in Real Analysis" (2012) by Protter & Morrey for a complete proof and further discussion.
Idea of proof
Let us consider a particular case of the method when
where
Since we know that
To proceed it is convenient to isolate the following result of linear algebra.
Consider
and let . If for all then for some .
In order to prove this we write
This means that
The above statement holds in any dimension with any number of vectors with the analogous proof. Applying this lemma to the vectors