Inverse function theorem

Inverse function theorem

In mathematics, specifically differential calculus, the inverse function theorem gives sufficient conditions for a function to be invertible in a neighborhood of a point in its domain. The theorem also gives a formula for the derivative of the inverse function.

In multivariable calculus, this theorem can be generalized to any vector-valued function whose Jacobian determinant is nonzero at a point in its domain. In this case, the theorem gives a formula for the Jacobian matrix of the inverse. There are also versions of the inverse function theorem for complex holomorphic functions, for differentiable maps between manifolds, for differentiable functions between Banach spaces, and so forth.

Contents

Statement of the theorem

For functions of a single variable, the theorem states that if ƒ is a continuously differentiable function with nonzero derivative at the point a, then ƒ is invertible in a neighborhood of a, the inverse is continuously differentiable, and

\bigl(f^{-1}\bigr)'(b) = \frac{1}{f'(a)}

where b = ƒ(a).

For functions of more than one variable, the theorem states that if the total derivative of a continuously differentiable function F defined from an open set U of Rn into Rn is invertible at a point p (i.e., the Jacobian determinant of F at p is non-zero), then F is an invertible function near p. That is, an inverse function to F exists in some neighborhood of F(p). Moreover, the inverse function F − 1 is also continuously differentiable. In the infinite dimensional case it is required that the Fréchet derivative have a bounded inverse at p.

Finally, the theorem says that

 J_{F^{-1}}(F(p)) = [ J_F(p) ]^{-1}

where [\cdot]^{-1} denotes matrix inverse and JG(q) is the Jacobian matrix of the function G at the point q.

This formula can also be derived from the chain rule. The chain rule states that for functions G and H which have total derivatives at H(p) and p respectively,

J_{G \circ H} (p) = J_G (H(p)) \cdot J_H (p).

Letting G be F -1 and H be F, G \circ H is the identity function, whose Jacobian matrix is also the identity. In this special case, the formula above can be solved for J_{F^{-1}}(F(p)). Note that the chain rule assumes the existence of total derivative of the inside function H, while the inverse function theorem proves that F-1 has a total derivative at p.

The existence of an inverse function to F is equivalent to saying that the system of n equations yi = Fj(x1,...,xn) can be solved for x1,...,xn in terms of y1,...,yn if we restrict x and y to small enough neighborhoods of p and F(p), respectively.

Example

Consider the vector-valued function F from R2 to R2 defined by


\mathbf{F}(x,y)=
\begin{bmatrix}
 {e^x \cos y}\\
 {e^x \sin y}\\
\end{bmatrix}.

Then the Jacobian matrix is


J_F(x,y)=
\begin{bmatrix}
 {e^x \cos y} & {-e^x \sin y}\\
 {e^x \sin y} & {e^x \cos y}\\
\end{bmatrix}

and the determinant is


\det J_F(x,y)=
e^{2x} \cos^2 y + e^{2x} \sin^2 y=
e^{2x}.
\,\!

The determinant e2x is nonzero everywhere. By the theorem, for every point p in R2, there exists a neighborhood about p over which F is invertible. Note that this is different than saying F is invertible over its entire image. In this example, F is not invertible because it is not injective (because f(x,y) = f(x,y + 2π).)

Notes on methods of proof

As an important result, the inverse function theorem has been given numerous proofs. The proof most commonly seen in textbooks relies on the contraction mapping principle, also known as the Banach fixed point theorem. (This theorem can also be used as the key step in the proof of existence and uniqueness of solutions to ordinary differential equations.) Since this theorem applies in infinite-dimensional (Banach space) settings, it is the tool used in proving the infinite-dimensional version of the inverse function theorem (see "Generalizations", below).

An alternate proof (which works only in finite dimensions) instead uses as the key tool the extreme value theorem for functions on a compact set.[1]

Yet another proof uses Newton's method, which has the advantage of providing an effective version of the theorem. That is, given specific bounds on the derivative of the function, an estimate of the size of the neighborhood on which the function is invertible can be obtained.[2]

Generalizations

Manifolds

The inverse function theorem can be generalized to differentiable maps between differentiable manifolds. In this context the theorem states that for a differentiable map F : MN, if the derivative of F,

(dF)p : TpM → TF(p)N

is a linear isomorphism at a point p in M then there exists an open neighborhood U of p such that

F|U : UF(U)

is a diffeomorphism. Note that this implies that M and N must have the same dimension at p.

If the derivative of F is an isomorphism at all points p in M then the map F is a local diffeomorphism.

Banach spaces

The inverse function theorem can also be generalized to differentiable maps between Banach spaces. Let X and Y be Banach spaces and U an open neighbourhood of the origin in X. Let F : U → Y be continuously differentiable and assume that the derivative (dF)0 : X → Y of F at 0 is a bounded linear isomorphism of X onto Y. Then there exists an open neighbourhood V of F(0) in Y and a continuously differentiable map G : V → X such that F(G(y)) = y for all y in V. Moreover, G(y) is the only sufficiently small solution x of the equation F(x) = y.

Banach manifolds

These two directions of generalization can be combined in the inverse function theorem for Banach manifolds.[3]

Constant rank theorem

The inverse function theorem (and the implicit function theorem) can be seen as a special case of the constant rank theorem, which states that a smooth map with locally constant rank near a point can be put in a particular normal form near that point.[4] When the derivative of F is invertible at a point p, it is also invertible in a neighborhood of p, and hence the rank of the derivative is constant, so the constant rank theorem applies.

See also

Notes

  1. ^ Michael Spivak, Calculus on Manifolds.
  2. ^ John H. Hubbard and Barbara Burke Hubbard, Vector Analysis, Linear Algebra, and Differential Forms: a unified approach, Matrix Editions, 2001.
  3. ^ Serge Lang, Differential and Riemannian Manifolds, Springer, 1995, ISBN 0387943382.
  4. ^ Wiilliam M. Boothby, An Introduction to Differentiable Manifolds and Riemannian Geometry, Academic Press, 2002, ISBN 0121160513.

References

  • Nijenhuis, Albert (1974). "Strong derivatives and inverse mappings". Amer. Math. Monthly 81 (9): 969–980. doi:10.2307/2319298. 
  • Renardy, Michael and Rogers, Robert C. (2004). An introduction to partial differential equations. Texts in Applied Mathematics 13 (Second ed.). New York: Springer-Verlag. pp. 337–338. ISBN 0-387-00444-0. 
  • Rudin, Walter (1976). Principles of mathematical analysis. International Series in Pure and Applied Mathematics (Third ed.). New York: McGraw-Hill Book Co.. pp. 221–223. 

Wikimedia Foundation. 2010.

Игры ⚽ Нужно сделать НИР?

Look at other dictionaries:

  • Inverse mapping theorem — In mathematics, inverse mapping theorem may refer to:* the inverse function theorem on the existence of local inverses for functions with non singular derivatives;* the bounded inverse theorem on the boundedness of the inverse for invertible… …   Wikipedia

  • Inverse function — In mathematics, if fnof; is a function from A to B then an inverse function for fnof; is a function in the opposite direction, from B to A , with the property that a round trip (a composition) from A to B to A (or from B to A to B ) returns each… …   Wikipedia

  • Implicit function theorem — In the branch of mathematics called multivariable calculus, the implicit function theorem is a tool which allows relations to be converted to functions. It does this by representing the relation as the graph of a function. There may not be a… …   Wikipedia

  • Inverse functions and differentiation — In mathematics, the inverse of a function y = f(x) is a function that, in some fashion, undoes the effect of f (see inverse function for a formal and detailed definition). The inverse of f is denoted f^{ 1}. The statements y=f(x) and x=f 1(y) are …   Wikipedia

  • Function (mathematics) — f(x) redirects here. For the band, see f(x) (band). Graph of example function, In mathematics, a function associates one quantity, the a …   Wikipedia

  • inverse — I. adjective Etymology: Middle English, turned upside down, from Latin inversus, from past participle of invertere Date: 15th century 1. opposite in order, nature, or effect 2. being an inverse function < inverse sine > II. noun Date: circa 1681 …   New Collegiate Dictionary

  • Nash–Moser theorem — The Nash–Moser theorem, attributed to mathematicians John Forbes Nash and Jürgen Moser, is a generalization of the inverse function theorem on Banach spaces to a class of tame Fréchet spaces. In contrast to the Banach space case, in which the… …   Wikipedia

  • Nash-Moser theorem — The Nash Moser theorem, attributed to mathematicians John Forbes Nash and Jurgen Moser is a generalization of the inverse function theorem on Banach spaces to a class of tame Frechet spaces.In contrast to the Banach space case, in which the… …   Wikipedia

  • Banach fixed point theorem — The Banach fixed point theorem (also known as the contraction mapping theorem or contraction mapping principle) is an important tool in the theory of metric spaces; it guarantees the existence and uniqueness of fixed points of certain self maps… …   Wikipedia

  • Banach fixed-point theorem — In mathematics, the Banach fixed point theorem (also known as the contraction mapping theorem or contraction mapping principle) is an important tool in the theory of metric spaces; it guarantees the existence and uniqueness of fixed points of… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”