Jacobian matrix: Difference between revisions
No edit summary |
|||
| (5 intermediate revisions by the same user not shown) | |||
| Line 11: | Line 11: | ||
===Definition at a point in terms of gradient vectors as row vectors=== | ===Definition at a point in terms of gradient vectors as row vectors=== | ||
Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots, | Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots,f_m(x_1,x_2,\dots,x_n)</math>. Suppose <math>(a_1,a_2,\dots,a_n)</math> is a point in the domain of <math>f</math> such that <math>f_i</math> is differentiable at <math>(a_1,a_2,\dots,a_n)</math> for <math>i \in \{ 1,2,\dots,m\}</math>. Then, the '''Jacobian matrix''' of <math>f</math> at <math>(a_1,a_2,\dots,a_n)</math> is a <math>m \times n</math> matrix of ''numbers'' whose <math>i^{th}</math> row is given by the [[gradient vector]] of <matH>f_i</math> at <math>(a_1,a_2,\dots,a_n)</math>. | ||
Explicitly, in terms of rows, it looks like: | |||
<math>\begin{pmatrix} \nabla(f_1)(a_1,a_2,\dots,a_n) \\ \nabla(f_2)(a_1,a_2,\dots,a_n) \\ \cdot \\ \cdot \\ \cdot \\ \nabla(f_m)(a_1,a_2,\dots,a_n) \\\end{pmatrix}</math> | |||
<center>{{#widget:YouTube|id=O8isoxng_9g}}</center> | |||
===Definition at a point in terms of partial derivatives=== | ===Definition at a point in terms of partial derivatives=== | ||
Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots, | Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots,f_m(x_1,x_2,\dots,x_n)</math>. Suppose <math>(a_1,a_2,\dots,a_n)</math> is a point in the domain of <math>f</math> such that <math>f_i</math> is differentiable at <math>(a_1,a_2,\dots,a_n)</math> for <math>i \in \{ 1,2,\dots,m\}</math>. Then, the '''Jacobian matrix''' of <math>f</math> at <math>(a_1,a_2,\dots,a_n)</math> is a <math>m \times n</math> matrix of ''numbers'' whose <math>(ij)^{th}</math> entry is given by: | ||
<math>\frac{\partial f_i}{\partial x_j}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)}</math> | <math>\frac{\partial f_i}{\partial x_j}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)}</math> | ||
Here's how the matrix looks: | |||
<math>\begin{pmatrix} \frac{\partial f_1}{\partial x_1}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} & \frac{\partial f_1}{\partial x_2}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} & \dots & \frac{\partial f_1}{\partial x_n}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} \\ | |||
\frac{\partial f_2}{\partial x_1}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} & \frac{\partial f_2}{\partial x_2}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} & \dots & \frac{\partial f_2}{\partial x_n}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} \\ | |||
\cdot & \cdot & \cdot & \cdot \\ | |||
\frac{\partial f_m}{\partial x_1}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} & \frac{\partial f_m}{\partial x_2}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} & \dots & \frac{\partial f_m}{\partial x_n}(x_1,x_2,\dots,x_n)|_{(x_1,x_2,\dots,x_n) = (a_1,a_2,\dots,a_n)} \\\end{pmatrix}</math> | |||
Note that for this definition to be correct, it is still necessary that the gradient vectors exist. If the gradient vectors do not exist but the partial derivatives do, a matrix can still be constructed using this recipe but it may not satisfy the nice behavior that the Jacobian matrix does. | Note that for this definition to be correct, it is still necessary that the gradient vectors exist. If the gradient vectors do not exist but the partial derivatives do, a matrix can still be constructed using this recipe but it may not satisfy the nice behavior that the Jacobian matrix does. | ||
<center>{{#widget:YouTube|id=VCM4RVM09_I}}</center> | |||
==Definition as a function== | ==Definition as a function== | ||
| Line 25: | Line 40: | ||
===Definition in terms of gradient vectors as row vectors=== | ===Definition in terms of gradient vectors as row vectors=== | ||
Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots, | Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots,f_m(x_1,x_2,\dots,x_n)</math>. Then, the '''Jacobian matrix''' of <math>f</math>is a <math>m \times n</math> matrix of ''functions'' whose <math>i^{th}</math> row is given by the [[gradient vector]] of <matH>f_i</math>. Explicitly, it looks like this: | ||
<math>\begin{pmatrix} \nabla(f_1) \\ \nabla(f_2)\\ \cdot \\ \cdot \\ \cdot \\ \nabla(f_m) \\\end{pmatrix}</math> | |||
Note that the domain of this function is the set of points at which all the <math>f_i</math>s individually are differentiable. | Note that the domain of this function is the set of points at which all the <math>f_i</math>s individually are differentiable. | ||
| Line 31: | Line 49: | ||
===Definition in terms of partial derivatives=== | ===Definition in terms of partial derivatives=== | ||
Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots, | Suppose <math>f</matH> is a vector-valued function with <math>n</math>-dimensional inputs and <math>m</math>-dimensional outputs. Explicitly, suppose <math>f</math> is a function with inputs <math>x_1,x_2,\dots,x_n</math> and outputs <math>f_1(x_1,x_2,\dots,x_n), f_2(x_1,x_2,\dots,x_n),\dots,f_m(x_1,x_2,\dots,x_n)</math>. Then, the '''Jacobian matrix''' of <math>f</math> is a <math>m \times n</math> matrix of ''functions'' whose <math>(ij)^{th}</math> entry is given by: | ||
<math>\frac{\partial f_i}{\partial x_j}(x_1,x_2,\dots,x_n)</math> | <math>\frac{\partial f_i}{\partial x_j}(x_1,x_2,\dots,x_n)</math> | ||
wherever all the <math>f_i</math>s individually are differentiable in the sense of the gradient vectors existing. If the gradient vectors do not exist but the partial derivatives do, a matrix can still be constructed using this recipe but it may not satisfy the nice behavior that the Jacobian matrix does. | wherever all the <math>f_i</math>s individually are differentiable in the sense of the gradient vectors existing. Here's how the matrix looks: | ||
<math>\begin{pmatrix} \frac{\partial f_1}{\partial x_1} & \frac{\partial f_1}{\partial x_2}& \dots & \frac{\partial f_1}{\partial x_n}\\ | |||
\frac{\partial f_2}{\partial x_1} & \frac{\partial f_2}{\partial x_2} & \dots & \frac{\partial f_2}{\partial x_n}\\ | |||
\cdot & \cdot & \cdot & \cdot \\ | |||
\frac{\partial f_m}{\partial x_1} & \frac{\partial f_m}{\partial x_2} & \dots & \frac{\partial f_m}{\partial x_n}\\\end{pmatrix}</math> | |||
If the gradient vectors do not exist but the partial derivatives do, a matrix can still be constructed using this recipe but it may not satisfy the nice behavior that the Jacobian matrix does. | |||
<center>{{#widget:YouTube|id=jTmwUMnuUec}}</center> | |||
==Particular cases== | ==Particular cases== | ||
Latest revision as of 02:32, 13 May 2012
This article describes an analogue for functions of multiple variables of the following term/fact/notion for functions of one variable: derivative
Importance
The Jacobian matrix is the appropriate notion of derivative for a function that has multiple inputs (or equivalently, vector-valued inputs) and multiple outputs (or equivalently, vector-valued outputs).
Definition at a point
Direct epsilon-delta definition
Definition at a point in terms of gradient vectors as row vectors
Suppose is a vector-valued function with -dimensional inputs and -dimensional outputs. Explicitly, suppose is a function with inputs and outputs . Suppose is a point in the domain of such that is differentiable at for . Then, the Jacobian matrix of at is a matrix of numbers whose row is given by the gradient vector of at .
Explicitly, in terms of rows, it looks like:
Definition at a point in terms of partial derivatives
Suppose is a vector-valued function with -dimensional inputs and -dimensional outputs. Explicitly, suppose is a function with inputs and outputs . Suppose is a point in the domain of such that is differentiable at for . Then, the Jacobian matrix of at is a matrix of numbers whose entry is given by:
Here's how the matrix looks:
Note that for this definition to be correct, it is still necessary that the gradient vectors exist. If the gradient vectors do not exist but the partial derivatives do, a matrix can still be constructed using this recipe but it may not satisfy the nice behavior that the Jacobian matrix does.
Definition as a function
Definition in terms of gradient vectors as row vectors
Suppose is a vector-valued function with -dimensional inputs and -dimensional outputs. Explicitly, suppose is a function with inputs and outputs . Then, the Jacobian matrix of is a matrix of functions whose row is given by the gradient vector of . Explicitly, it looks like this:
Note that the domain of this function is the set of points at which all the s individually are differentiable.
Definition in terms of partial derivatives
Suppose is a vector-valued function with -dimensional inputs and -dimensional outputs. Explicitly, suppose is a function with inputs and outputs . Then, the Jacobian matrix of is a matrix of functions whose entry is given by:
wherever all the s individually are differentiable in the sense of the gradient vectors existing. Here's how the matrix looks:
If the gradient vectors do not exist but the partial derivatives do, a matrix can still be constructed using this recipe but it may not satisfy the nice behavior that the Jacobian matrix does.
Particular cases
| Case | What happens in that case? |
|---|---|
| is a real-valued function of one variable. The Jacobian matrix is a matrix whose entry is the ordinary derivative. | |
| , | is a vector-valued function of one variable. We can think of it as a parametric curve in . The Jacobian matrix is a matrix which, read as a column vector, is the parametric derivative of the vector-valued function. |
| , | is a real-valued function of multiple variables. The Jacobian matrix is a matrix which, read as a row vector, is the gradient vector function. |
| is a linear or affine map. | The Jacobian matrix is the same as the matrix describing (or, if is affine, the matrix describing the linear part of ). |
| , and we are identifying the spaces of inputs and outputs of . | The Jacobian matrix can then be thought of as a linear self-map from the -dimensional space to itself. In this context, we can consider the Jacobian determinant. |