Calculus of variations is a method for proving existence and uniqueness results for certain equations; in particular, it can be applied to some partial differential equations. The method works as follows: Let's say we have an equation which is to be solved for the variable
(this variable can also be a function). We look for a function whose minimizers satisfy the equation, and then prove that there exists a minimizer. We have thus obtained an existence result.
In some cases, we will additionally be able to show that values
satisfying the equation are minimizers of the function. If we now find out about the number of minimizers of the function, we will also know the numbers of solutions to the equation. If then the function has only one minimizer, we have obtained a uniqueness result.
Sometimes, calculus of variations also works ‘the other way round’: We have a function whose minimizers are difficult to find. Then we show that the minimizers of this function are exactly the solutions of a partial differential equation, which is easy to solve. We then solve the partial differential equation in order to obtain the minimizers of the function.
Consider the equation system

for functions
. If there exists a function
such that

we find that the equation system
is satisfied if and only if

If
satisfies the right conditions, we have
at exactly one point
:
Proof:
From
being strongly convex it follows that for all
,
is positively definite. Therefore, every critical point is a local minimum (this is due to the sufficient condition for local minima). Thus, it suffices to prove that there is exactly one local minimum.
1.
We show that there exists a local minimum.
We take Taylor's formula around
:
![{\displaystyle \forall x\in \mathbb {R} ^{d}:\exists \lambda \in [0,1]:f(x)=f(0)+x^{T}\nabla f(0)+{\frac {1}{2}}x^{T}H_{f}(\lambda x)x}](https://wikimedia.org/api/rest_v1/media/math/render/svg/ff4b82286e7eb92c4299dc64489447d4962039bc)
Thus,
![{\displaystyle {\begin{aligned}\forall x\in \mathbb {R} ^{d}:f(x)&=f(0)+x^{T}\nabla f(0)+{\frac {1}{2}}x^{T}H_{f}(\lambda x)x&{\text{ for a }}\lambda \in [0,1]\\&\geq f(0)+x^{T}\nabla f(0)+{\frac {c}{2}}\|x\|^{2}&f{\text{ is strongly convex}}\\&\geq f(0)-\|x\|\|\nabla f(0)\|+{\frac {c}{2}}\|x\|^{2}&{\text{Cauchy-Schwarz inequality}}\end{aligned}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/cd56732ff66aeb96145f80ee6e9b6e63f1364154)
for a
. Therefore, there exists an
such that

By the extreme value theorem, there exists a minimum
of
in
. It can not be attained on the border, because if
, then
and thus by
, which would imply that
is not a minimum. Therefore it is attained in the interior and is thus a local minimum. In fact, from
and from
being a minimum on
even follows that it is a global minimum of
.
2.
We show that there is only one local minimum.
Let
and
be two local minima. We show that
, thereby excluding the possibility of two different minima. We define a function
as follows:

Let's calculate the first and second derivative of
:

Since
and
are local minima,
and
. Therefore,

and

Therefore, by the mean value theorem, there exists a
such that

But since

,
implies
.
Corollary 13.3:
Suppose we have an equation system

If there is a function
which is strongly convex and

, then the equation system
has exactly one solution.
Proof: See exercise 1.
Example 13.4:
Another example is given in exercise 2.