# Elliptic boundary value problem

﻿
Elliptic boundary value problem

In mathematics, an elliptic boundary value problem is a special kind of boundary value problem which can be thought of as the stable state of an evolution problem. For example, the Dirichlet problem for the Laplacian gives the eventual distribution of heat in a room several hours after the heating is turned on.

Differential equations describe a large class of natural phenomena, from the heat equation describing the evolution of heat in (for instance) a metal plate, to the Navier-Stokes equation describing the movement of fluids, including Einstein's equations describing the physical universe in a relativistic way. Although all these equations are boundary value problems, they are further subdivided into categories. This is necessary because each category must be analyzed using different techniques. The present article deals with the category of boundary value problems known as linear elliptic problems.

Boundary value problems and partial differential equations specify relations between two or more quantities. For instance, in the heat equation, the rate of change of temperature at a point is related to the difference of temperature between that point and the nearby points so that, over time, the heat flows from hotter points to cooler points. Boundary value problems can involve space, time and other quantities such as temperature, velocity, pressure, magnetic field, etc...

Some problems do not involve time. For instance, if one hangs a clothesline between the house and a tree, then in the absence of wind, the clothesline will not move and will adopt a gentle hanging curved shape known as the catenary [Swetz, Faauvel, Bekken, "Learn from the Masters", 1997, MAA ISBN 0883857030, pp.128-9] . This curved shape can be computed as the solution of a differential equation relating position, tension, angle and gravity, but since the shape does not change over time, there is no time variable.

Elliptic boundary value problems are a class of problems which do not involve the time variable, and instead only depend on space variables.

It is not possible to discuss elliptic boundary value problems in more detail without referring to calculus in multiple variables.

Unless otherwise noted, all facts presented in this article can be found in [Partial Differential Equations by Lawrence C. Evans. American Mathematical Society, Providence, RI, 1998. Graduate Studies in Mathematics 19.] .

The main example

In two dimensions, let $x,y$ be the coordinates. We will use the notation $u_x, u_\left\{xx\right\}$ for the first and second partial derivatives of $u$ with respect to $x$, and a similar notation for $y$. We will use the symbols $D_x$ and $D_y$ for the partial derivatives in $x$ and $y$. The second partial derivatives will be denoted $D_x^2$ and $D_y^2$. We also define the gradient $abla u = \left(u_x,u_y\right)$, the Laplacian $Delta u = u_\left\{xx\right\}+u_\left\{yy\right\}$ and the divergence $abla cdot \left(u,v\right) = u_x + v_y$. Note from the definitions that $Delta u = abla cdot \left( abla u\right)$.

The main example for boundary value problems is the Laplacian,

:$Delta u = f ext\left\{ in \right\}Omega$,:$u = 0 ext \left\{ on \right\}partial Omega$;

where $Omega$ is a region in the plane and $partial Omega$ is the boundary of that region. The function $f$ is known data and the solution $u$ is what must be computed. This example has the same essential properties as all other elliptic boundary value problems.

The solution $u$ can be interpreted as the stationary or limit distribution of heat in a metal plate shaped like $Omega$, if this metal plate has its boundary adjacent to ice (which is kept at zero degrees, thus the Dirichlet condition.) The function $f$ represents the intensity of heat generation at each point in the plate (perhaps there is an electric heater resting on the metal plate, pumping heat into the plate at rate $f\left(x\right)$, which does not vary over time, but may be nonuniform in space on the metal plate.) After waiting for a long time, the temperature distribution in the metal plate will approach $u$.

Nomenclature

Let $Lu=a u_\left\{xx\right\} + b u_\left\{yy\right\}$ where $a$ and $b$ are constants. $L=aD_x^2+bD_y^2$ is called a second order differential operator. If we formally replace the derivatives $D_x$ by $x$ and $D_y$ by $y$, we obtain the expression

:$a x^2 + b y^2$.

If we set this expression equal to some constant $k$, then we obtain either an ellipse (if $a,b,k$ are all the same sign) or a hyperbola (if $a$ and $b$ are of opposite signs.) For that reason, $L$ is said to be elliptic when $ab>0$ and hyperbolic if $ab<0$. Similarly, the operator $L=D_x+D_y^2$ leads to a parabola, and so this $L$ is said to be parabolic.

We now generalize the notion of ellipticity. While it may not be obvious that our generalization is the right one, it turns out that it does preserve most of the necessary properties for the purpose of analysis.

General linear elliptic boundary value problems of the second degree

Let $x_1,...,x_n$ be the space variables. Let $a_\left\{ij\right\}\left(x\right), b_i\left(x\right), c\left(x\right)$ be real valued functions of $x=\left(x_1,...,x_n\right)$. Let $L$ be a second degree linear operator. That is,

:$Lu\left(x\right)=sum_\left\{i,j=1\right\}^n \left(a_\left\{ij\right\} \left(x\right) u_\left\{x_i\right\}\right)_\left\{x_j\right\} + sum_\left\{i=1\right\}^n b_i\left(x\right) u_\left\{x_i\right\}\left(x\right) + c\left(x\right) u\left(x\right)$ (divergence form).:$Lu\left(x\right)=sum_\left\{i,j=1\right\}^n a_\left\{ij\right\} \left(x\right) u_\left\{x_i x_j\right\} + sum_\left\{i=1\right\}^n ilde b_i u_\left\{x_i\right\}\left(x\right) + c\left(x\right) u\left(x\right)$ (nondivergence form)

We have used the subscript $cdot_\left\{x_i\right\}$ to denote the partial derivative with respect to the space variable $x_i$. The two formulae are equivalent, provided that

:$ilde b_i\left(x\right) = b_i\left(x\right) + sum_j a_\left\{ij,x_j\right\}\left(x\right)$.

In matrix notation, we can let $a\left(x\right)$ be an $n imes n$ matrix valued function of $x$ and $b\left(x\right)$ be a $n$-dimensional column vector-valued function of $x$, and then we may write

:$Lu = abla cdot \left(a abla u\right) + b^T abla u + c u$ (divergence form).

One may assume, without loss of generality, that the matrix $a$ is symmetric (that is, for all $i,j,x$, $a_\left\{ij\right\}\left(x\right)=a_\left\{ji\right\}\left(x\right)$. We make that assumption in the rest of this article.

We say that the operator $L$ is "elliptic" if, for some constant $alpha>0$, any of the following equivalent conditions hold:

# $lambda_\left\{min\right\} \left(a\left(x\right)\right) > alpha ;;; forall x$ (see eigenvalue).
# $u^T a\left(x\right) u > alpha u^T u ;;; forall u in mathbb\left\{R\right\}^n$.
# $sum_\left\{i,j=1\right\}^n a_\left\{ij\right\} u_i u_j > alpha sum_\left\{i=1\right\}^n u_i^2 ;;; forall u in mathbb\left\{R\right\}^n$.

An elliptic boundary value problem is then a system of equations like

:$Lu=f ext\left\{ in \right\} Omega$ (the PDE) and:$u=0 ext\left\{ on \right\} partial Omega$ (the boundary value).

This particular example is the Dirichlet problem. The Neumann problem is

:$Lu=f ext\left\{ in \right\} Omega$ and:$u_ u = g ext\left\{ on \right\} partial Omega$

where $u_ u$ is the derivative of $u$ in the direction of the outwards pointing normal of $partial Omega$. In general, if $B$ is any trace operator, one can construct the boundary value problem

:$Lu=f ext\left\{ in \right\} Omega$ and:$Bu=g ext\left\{ on \right\} partial Omega$.

In the rest of this article, we assume that $L$ is elliptic and that the boundary condition is the Dirichlet condition $u=0 ext\left\{ on \right\}partial Omega$.

Sobolev spaces

The analysis of elliptic boundary value problems requires some fairly sophisticated tools of functional analysis. We require the space $H^1\left(Omega\right)$, the Sobolev space of "once-differentiable" functions on $Omega$, such that both the function $u$ and its partial derivatives $u_\left\{x_i\right\}$, $i=1,dots,n$ are all square integrable. There is a subtlety here in that the partial derivatives must be defined "in the weak sense" (see the article on Sobolev spaces for details.) The space $H^1$ is a Hilbert space, which accounts for much of the ease with which these problems are analyzed.

The discussion in details of Sobolev spaces is beyond the scope of this article, but we will quote required results as they arise.

Unless otherwise noted, all derivatives in this article are to be interpreted in the weak, Sobolev sense. We use the term "strong derivative" to refer to the classical derivative of calculus. We also specify that the spaces $C^k$, $k=0,1,dots$ consist of functions that are $k$ times strongly differentiable, and that the $k$th derivative is continuous.

Weak or variational formulation

The first step to cast the boundary value problem as in the language of Sobolev spaces is to rephrase it in its weak form. Consider the Laplace problem $Delta u = f$. Multiply each side of the equation by a "test function" $varphi$ and integrate by parts using Green's theorem to obtain

:$-int_Omega abla u cdot abla varphi + int_\left\{partial Omega\right\} u_ u varphi = int_Omega f varphi$.

We will be solving the Dirichlet problem, so that $u=0 ext\left\{ on \right\}partial Omega$. For technical reasons, it is useful to assume that $varphi$ is taken from the same space of functions as $u$ is so we also assume that $varphi=0 ext\left\{ on \right\}partial Omega$. This gets rid of the $int_\left\{partial Omega\right\}$ term, yielding

:$A\left(u,varphi\right) = F\left(varphi\right)$ (*)

where

:$A\left(u,varphi\right) = int_Omega abla u cdot abla varphi$ and:$F\left(varphi\right) = -int_Omega f varphi$.

If $L$ is a general elliptic operator, the same reasoning leads to the bilinear form

:$A\left(u,varphi\right) = int_Omega abla u ^T a abla varphi - int_Omega b^T abla u varphi - int_Omega c u varphi$.

We do not discuss the Neumann problem but note that it is analyzed in a similar way.

Continuous and coercive bilinear forms

The map $A\left(u,varphi\right)$ is defined on the Sobolev space $H^1_0subset H^1$ of functions which are once differentiable and zero on the boundary $partial Omega$, provided we impose some conditions on $a,b,c$ and $Omega$. There are many possible choices, but for the purpose of this article, we will assume that

# $a_\left\{ij\right\}\left(x\right)$ is continuously differentiable on for $i,j=1,dots,n$,
# $b_i\left(x\right)$ is continuous on for $i=1,dots,n$,
# $c\left(x\right)$ is continuous on and
# $Omega$ is bounded.

The reader may verify that the map $A\left(u,varphi\right)$ is furthermore bilinear and continuous, and that the map $F\left(varphi\right)$ is linear in $varphi$, and continuous if (for instance) $f$ is square integrable.

We say that the map $A$ is coercive if there is an $alpha>0$ for all $u,varphi in H_0^1\left(Omega\right)$,

:$A\left(u,varphi\right) geq alpha int_Omega abla u cdot abla varphi$.

This is trivially true for the Laplacian (with $alpha=1$) and is also true for an elliptic operator if we assume $b = 0$ and $c leq 0$. (Recall that $u^T a u > alpha u^T u$ when $L$ is elliptic.)

Existence and uniqueness of the weak solution

One may show, via the Lax-Milgram lemma, that whenever $A\left(u,varphi\right)$ is coercive and $F\left(varphi\right)$ is continuous, then there exists a unique solution $uin H_0^1\left(Omega\right)$ to the weak problem (*).

If further $A\left(u,varphi\right)$ is symmetric (i.e., $b=0$), one can show the same result using the Riesz representation theorem instead.

This relies on the fact that $A\left(u,varphi\right)$ forms an inner product on $H_0^1\left(Omega\right)$, which itself depends on Poincaré's inequality.

Strong solutions

We have shown that there is a $uin H_0^1\left(Omega\right)$ which solves the weak system, but we do not know if this $u$ solves the strong system

:$Lu=f ext\left\{ in \right\}Omega$,:$u=0 ext\left\{ on \right\}partial Omega$.

Even more vexing is that we are not even sure that $u$ is twice differentiable, rendering the expressions $u_\left\{x_i x_j\right\}$ in $Lu$ apparently meaningless. There are many ways to remedy the situation, the main one being regularity.

Regularity

A regularity theorem for a linear elliptic boundary value problem of the second order takes the form

Theorem "If (some condition), then the solution $u$ is in $H^2\left(Omega\right)$, the space of "twice differentiable" functions whose second derivatives are square integrable."

There is no known simple condition necessary and sufficient for the conclusion of the theorem to hold, but the following conditions are known to be sufficient:

# The boundary of $Omega$ is $C^2$, or
# $Omega$ is convex.

It may be tempting to infer that if $partial Omega$ is piecewise $C^2$ then $u$ is indeed in $H^2$, but that is unfortunately false.

Almost everywhere solutions

In the case that $u in H^2\left(Omega\right)$ then the second derivatives of $u$ are defined almost everywhere, and in that case $Lu=f$ almost everywhere.

Strong solutions

One may further prove that if the boundary of $Omega subset mathbb\left\{R\right\}^n$ is a smooth manifold and $f$ is infinitely differentiable in the strong sense, then $u$ is also infinitely differentiable in the strong sense. In this case, $Lu=f$ with the strong definition of the derivative.

The proof of this relies upon an improved regularity theorem that says that if $partial Omega$ is $C^k$ and $f in H^\left\{k-2\right\}\left(Omega\right)$, $kgeq 2$, then $uin H^k\left(Omega\right)$, together with a Sobolev imbedding theorem saying that functions in $H^k\left(Omega\right)$ are also in whenever $0 leq m < k-n/2$.

Numerical solutions

While in exceptional circumstances, it is possible to solve elliptic problems explicitly, in general it is an impossible task. The natural solution is to approximate the elliptic problem with a simpler one and to solve this simpler problem on a computer.

Because of the good properties we have enumerated (as well as many we have not), there are extremely efficient numerical solvers for linear elliptic boundary value problems (see finite element method, finite difference method and spectral method for examples.)

Eigenvalues and eigensolutions

Another Sobolev imbedding theorem states that the inclusion $H^1subset L^2$ is a compact linear map. Equipped with the spectral theorem for compact linear operators, one obtains the following result.

Theorem "Assume that $A\left(u,varphi\right)$ is coercive, continuous and symmetric. The map $S : f ightarrow u$ from $L^2\left(Omega\right)$ to $L^2\left(Omega\right)$ is a compact linear map. It has a basis of eigenvectors $u_1, u_2, dots in H^1\left(Omega\right)$ and matching eigenvalues $lambda_1,lambda_2,dots in mathbb\left\{R\right\}$ such that"

# $Su_k = lambda_k u_k, k=1,2,dots,$
# $lambda_k ightarrow 0$ "as" $k ightarrow infty$,
# $lambda_k gneqq 0;;forall k$,
# $int_\left\{Omega\right\} u_j u_k = 0$ "whenever" $j eq k$ "and"
# $int_\left\{Omega\right\} u_j u_j = 1$ "for all" $j=1,2,dots,.$

Series solutions and the importance of eigensolutions

If one has computed the eigenvalues and eigenvectors, then one may find the "explicit" solution of $Lu=f$,

:$u=sum_\left\{k=1\right\}^infty hat u\left(k\right) u_k$

via the formula

:$hat u\left(k\right) = lambda_k hat f\left(k\right) ,;;k=1,2,dots$

where

:$hat f\left(k\right) = int_\left\{Omega\right\} f\left(x\right) u_k\left(x\right) , dx$.

(See Fourier series.)

The series converges in $L^2$. Implemented on a computer using numerical approximations, this is known as the spectral method.

An example

Consider the problem

:$u-u_\left\{xx\right\}-u_\left\{yy\right\}=f\left(x,y\right)=xy$ on $\left(0,1\right) imes\left(0,1\right)$,:$u\left(x,0\right)=u\left(x,1\right)=u\left(0,y\right)=u\left(1,y\right)=0 ;;forall \left(x,y\right)in\left(0,1\right) imes\left(0,1\right)$ (Dirichlet conditions).

The reader may verify that the eigenvectors are exactly

:$u_\left\{jk\right\}\left(x,y\right)=sin\left(pi jx\right)sin\left(pi ky\right)$, $j,kin mathbb\left\{N\right\}$

with eigenvalues

:$lambda_\left\{jk\right\}=\left\{ 1 over 1+pi^2 j^2+pi^2 k^2 \right\}$.

The Fourier coefficients of $g\left(x\right)=x$ can be looked up in a table, getting $hat g\left(n\right) = \left\{ \left(-1\right)^\left\{n+1\right\} over pi n \right\}$. Therefore,

:$hat f\left(j,k\right) = \left\{ \left(-1\right)^\left\{j+k+1\right\} over pi^2 jk \right\}$

yielding the solution

:$u\left(x,y\right) = sum_\left\{j,k=1\right\}^infty \left\{ \left(-1\right)^\left\{j+k+1\right\} over pi^2 jk \left(1+pi^2 j^2+pi^2 k^2\right) \right\} sin\left(pi jx\right) sin \left(pi ky\right)$.

Maximum principle

There are many variants of the maximum principle. We give a simple one.

Theorem. "(Weak maximum principle.) Let , and assume that $c\left(x\right)=0;forall xinOmega$. Say that $Lu leq 0$ in $Omega$. Then . In other words, the maximum is attained on the boundary."

A strong maximum principle would conclude that $u\left(x\right) lneqq max_\left\{y in partial Omega\right\} u\left(y\right)$ for all $x in Omega$ unless $u$ is constant.

References

Wikimedia Foundation. 2010.

### Look at other dictionaries:

• Boundary value problem — In mathematics, in the field of differential equations, a boundary value problem is a differential equation together with a set of additional restraints, called the boundary conditions. A solution to a boundary value problem is a solution to the… …   Wikipedia

• Stochastic processes and boundary value problems — In mathematics, some boundary value problems can be solved using the methods of stochastic analysis. Perhaps the most celebrated example is Shizuo Kakutani s 1944 solution of the Dirichlet problem for the Laplace operator using Brownian motion.… …   Wikipedia

• Examples of boundary value problems — We will use k to denote the square root of the absolute value of lambda.If lambda = 0 then:y(x) = Ax + B,solves the ODE. Substituted boundary conditions give that both A and B are equal to zero.For positive lambda we obtain that:y(x) = A e^{kx} + …   Wikipedia

• Obstacle problem — The obstacle problem is a classic motivating example in the mathematical study of variational inequalities and free boundary problems. The problem is to find the equilibrium position of an elastic membrane whose boundary is held fixed, and which… …   Wikipedia

• Finite element method — The finite element method (FEM) (sometimes referred to as finite element analysis) is a numerical technique for finding approximate solutions of partial differential equations (PDE) as well as of integral equations. The solution approach is based …   Wikipedia

• List of mathematics articles (E) — NOTOC E E₇ E (mathematical constant) E function E₈ lattice E₈ manifold E∞ operad E7½ E8 investigation tool Earley parser Early stopping Earnshaw s theorem Earth mover s distance East Journal on Approximations Eastern Arabic numerals Easton s… …   Wikipedia

• Compact operator — In functional analysis, a branch of mathematics, a compact operator is a linear operator L from a Banach space X to another Banach space Y, such that the image under L of any bounded subset of X is a relatively compact subset of Y. Such an… …   Wikipedia

• Séminaire Nicolas Bourbaki (1950–1959) — Continuation of the Séminaire Nicolas Bourbaki programme, for the 1950s. 1950/51 series *33 Armand Borel, Sous groupes compacts maximaux des groupes de Lie, d après Cartan, Iwasawa et Mostow (maximal compact subgroups) *34 Henri Cartan, Espaces… …   Wikipedia

• Séminaire Nicolas Bourbaki (1960–1969) — Continuation of the Séminaire Nicolas Bourbaki programme, for the 1960s.1960/61 series*205 Adrien Douady, Plongements de sphères, d après Mazur et Brown (embeddings of spheres) *206 Roger Godement, Groupes linéaires algébriques sur un corps… …   Wikipedia

• Hilbert-Schmidt theorem — In mathematical analysis, the Hilbert Schmidt theorem, also known as the eigenfunction expansion theorem, is a fundamental result concerning compact, self adjoint operators on Hilbert spaces. In the theory of partial differential equations, it is …   Wikipedia