Hessian matrix: Difference between revisions

From Calculus
Line 3: Line 3:
===For a function of two variables at a point===
===For a function of two variables at a point===


Suppose <math>f</math> is a real-valued function of two variables <math>x,y</math> and <math>(x_0,y_0)</math> is a point in the domain of <math>f</math>. Suppose all the four second-order partial derivatives exist at <math>(x_0,y_0)</math>, i.e., the two pure second-order partials <math>f_{xx}(x_0,y_0),f_{yy}(x_0,y_0)</math> exist, and so do the two [[second-order mixed partial derivative]]s <math>f_{xy}(x_0,y_0</math> and <math>f_{yx}(x_0,y_0)</math>. Then, the Hessian matrix of <math>f</math> at <math>(x_0,y_0)</math>, denoted <math>H(f)(x_0,y_0)</math>, is a <math>2 \times 2</math> matrix of real numbers defined as follows:
Suppose <math>f</math> is a real-valued function of two variables <math>x,y</math> and <math>(x_0,y_0)</math> is a point in the domain of <math>f</math>. Suppose all the four second-order partial derivatives exist at <math>(x_0,y_0)</math>, i.e., the two pure second-order partials <math>f_{xx}(x_0,y_0),f_{yy}(x_0,y_0)</math> exist, and so do the two [[second-order mixed partial derivative]]s <math>f_{xy}(x_0,y_0)</math> and <math>f_{yx}(x_0,y_0)</math>. Then, the Hessian matrix of <math>f</math> at <math>(x_0,y_0)</math>, denoted <math>H(f)(x_0,y_0)</math>, is a <math>2 \times 2</math> matrix of real numbers defined as follows:


<math>\begin{pmatrix} f_{xx}(x_0,y_0) & f_{xy}(x_0,y_0) \\ f_{yx}(x_0,y_0) & f_{yy}(x_0,y_0) \\\end{pmatrix}</math>
<math>\begin{pmatrix} f_{xx}(x_0,y_0) & f_{xy}(x_0,y_0) \\ f_{yx}(x_0,y_0) & f_{yy}(x_0,y_0) \\\end{pmatrix}</math>

Revision as of 00:59, 24 April 2012

Definition at a point

For a function of two variables at a point

Suppose is a real-valued function of two variables and is a point in the domain of . Suppose all the four second-order partial derivatives exist at , i.e., the two pure second-order partials exist, and so do the two second-order mixed partial derivatives and . Then, the Hessian matrix of at , denoted , is a matrix of real numbers defined as follows:

For a function of multiple variables at a point

Suppose is a real-valued function of multiple variables . Suppose is a point in the domain of . In other words, are real numbers and the point has coordinates . Suppose, further, that all the second-order partials (pure and mixed) of with respect to these variables exist at the point . Then, the Hessian matrix of at , denoted , is a matrix of real numbers defined as follows:

The entry (i.e., the entry in the row and column) is . This is the same as . Note that in the two notations, the order in which we write the partials differs because the convention differs (left-to-right versus right-to-left).

The matrix looks like this:

Definition as a function

For a function of two variables

Suppose is a real-valued function of two variables . The Hessian matrix of , denoted , is a matrix-valued function that sends each point to the Hessian matrix at that point, if that matrix is defined. It is defined as:

In the point-free notation, we can write this as:

For a function of multiple variables

Suppose is a function of variables . The Hessian matrix of , denoted , is a matrix-valued function that sends each point to the Hessian matrix at that point, if the matrix is defined. It is defined as:

In the point-free notation, we can write it as:

Under continuity assumptions

If we assume that all the second-order partials of are continuous functions everywhere, then the following happens:

  • The Hessian matrix of at any point is a symmetric matrix, i.e., its entry equals its entry. This follows from Clairaut's theorem on equality of mixed partials.
  • We can think of the Hessian matrix as the second derivative of the function, i.e., it is a matrix describing the second derivative.
  • is twice differentiable as a function. Hence, the Hessian matrix of is the same as the Jacobian matrix of the gradient vector , where the latter is viewed as a vector-valued function.

Note that the final conclusion actually only requires the existence of the gradient vector, hence it holds even if the second-order partials are not continuous.