Quadratic function of multiple variables: Difference between revisions
| Line 32: | Line 32: | ||
<math>\frac{\partial f}{\partial x_i} = \left(\sum_{j=1}^n a_{ij}x_j\right) + b_i</math> | <math>\frac{\partial f}{\partial x_i} = \left(\sum_{j=1}^n a_{ij}x_j\right) + b_i</math> | ||
In terms of the matrix and vector notation, the gradient vector, expressed as a column vector, is: | |||
<math>(\nabla f)(\vec{x}) = A\vec{x} + \vec{b}</math> | |||
This can be obtained by applying the product rule for differentiation to the functional form. | |||
===Hessian matrix=== | ===Hessian matrix=== | ||
Revision as of 18:30, 11 May 2014
Definition
Consider variables . A quadratic function of the variables is a function of the form:
In vector form, if we denote by the column vector with coordinates , then we can write the function as:
where is the matrix with entries and is the column vector with entries .
Key data
| Item | Value |
|---|---|
| default domain | the whole of |
| range | If the matrix is not positive semidefinite or negative semidefinite, the range is all of . If the matrix is positive semidefinite, the range is where is the minimum value. If the matrix is negative semidefinite, the range is where is the maximum value. |
| local minimum value and points of attainment | If the matrix is positive definite, then , attained at (also applies if it's positive semidefinite) Otherwise, no local minimum value |
| local maximum value and points of attainment | If the matrix is negative definite, then , attained at (also applies if it's negative semidefinite) Otherwise, no local maximum value |
Differentiation
Partial derivatives and gradient vector
The partial derivative with respect to the variable , and therefore also the coordinate of the gradient vector, is given by:
In terms of the matrix and vector notation, the gradient vector, expressed as a column vector, is:
This can be obtained by applying the product rule for differentiation to the functional form.
Hessian matrix
The Hessian matrix of the quadratic function is the matrix .
Higher derivatives
All the higher derivative tensors are zero.
Cases
Positive definite case
First, we consider the case where is a positive definite matrix. In other words, we can write in the form:
where is a invertible matrix.
We can "complete the square" for this function:
In other words:
This is minimized when the expression whose norm we are measuring is zero, so that it is minimized when we have:
Simplifying, we obtain that we minimum occurs at:
Moreover, the value of the minimum is: