Comparison of vector algebra and geometric algebra

Comparison of vector algebra and geometric algebra

Vector algebra and geometric algebra are alternative approaches to providing additional algebraic structures on vector spaces, with geometric interpretations, particularly vector fields in multivariable calculus and applications in mathematical physics.

Vector algebra is simpler, but specific to Euclidean 3-space, while geometric algebra uses multilinear algebra, but works in all dimensions and signatures, notably 3+1 spacetime, as well as 2 dimensions. They are mathematically equivalent in 3 dimensions, though the approaches differ. Vector algebra is more widely used in elementary multivariable calculus, while geometric algebra is used in some more advanced treatments, and is proposed for elementary use as well. In advanced mathematics, particularly differential geometry, neither is widely used, with differential forms being far more widely used.

Contents

Basic concepts and operations

In vector algebra the basic objects are scalars and vectors, and the operations (beyond the vector space operations of scalar multiplication and vector addition) are the dot (or scalar) product and the cross product ×.

In geometric algebra the basic objects are multivectors (scalars are 0-vectors, vectors are 1-vectors, etc.), and the operations include the Clifford product (here called "geometric product") and the exterior product. The dot product/inner product/scalar product is defined on 1-vectors, and allows the geometric product to be expressed as the sum of the inner product and the exterior product when multiplying 1-vectors.

A distinguishing feature is that vector algebra uses the cross product, while geometric algebra uses the exterior product (and the geometric product). More subtly, geometric algebra in Euclidean 3-space distinguishes 0-vectors, 1-vectors, 2-vectors, and 3-vectors, while elementary vector algebra identifies 1-vectors and 2-vectors (as vectors) and 0-vectors and 3-vectors (as scalars), though more advanced vector algebra distinguishes these as scalars, vectors, pseudovectors, and pseudoscalars. Unlike vector algebra, geometric algebra includes sums of k-vectors of differing k.

The cross product does not generalize to dimensions other than 3 (as a product of two vectors, yielding a third vector), and in higher dimensions not all k-vectors can be identified with vectors or scalars. By contrast, the exterior product (and geometric product) is defined uniformly for all dimensions and signatures, and multivectors are closed under these operations.

Embellishments, ad hoc techniques, and tricks

More advanced treatments of vector algebra add embellishments to the initial picture – pseudovectors and pseudoscalars (in geometric algebra terms, 2-vectors and 3-vectors), while applications to other dimensions use ad hoc techniques and "tricks" rather than a general mathematical approach. By contrast, geometric algebra begins with a complete picture, and applies uniformly in all dimensions.

For example, applying vector calculus in 2 dimensions, such as to compute torque or curl, requires adding an artificial 3rd dimension and extending the vector field to be constant in that dimension. The torque or curl is then a normal vector field in this 3rd dimension. By contrast, geometric algebra in 2 dimensions defines torque and curl as pseudoscalar fields (2-vector fields), without requiring a 3rd dimension. Similarly, the scalar triple product is ad hoc, and can instead be expressed uniformly using the exterior product and the geometric product.

List of analogous formulas

Here are some comparisons between standard {\mathbb R}^3 vector relations and their corresponding wedge product and geometric product equivalents. All the wedge and geometric product equivalents here are good for more than three dimensions, and some also for two. In two dimensions the cross product is undefined even if what it describes (like torque) is perfectly well defined in a plane without introducing an arbitrary normal vector outside of the space.

Many of these relationships only require the introduction of the wedge product to generalize, but since that may not be familiar to somebody with only a traditional background in vector algebra and calculus, some examples are given.

Algebraic and geometric properties of cross and wedge products

Cross and wedge products are both antisymmetric:

\mathbf v \times \mathbf u = - (\mathbf u \times \mathbf v)
\mathbf v \wedge \mathbf u = - (\mathbf u \wedge \mathbf v)

They are both linear in the first operand

(\mathbf u + \mathbf v) \times \mathbf w = \mathbf u \times \mathbf w + \mathbf v \times \mathbf w
(\mathbf u + \mathbf v) \wedge \mathbf w = \mathbf u \wedge \mathbf w + \mathbf v \wedge \mathbf w

and in the second operand

\mathbf u \times (\mathbf v + \mathbf w)= \mathbf u \times \mathbf v + \mathbf u \times \mathbf w
\mathbf u \wedge (\mathbf v + \mathbf w)= \mathbf u \wedge \mathbf v + \mathbf u \wedge \mathbf w

In general, the cross product is not associative, while the wedge product is

(\mathbf u \times \mathbf v) \times \mathbf w \neq \mathbf u \times (\mathbf v \times \mathbf w)
(\mathbf u \wedge \mathbf v) \wedge \mathbf w = \mathbf u \wedge (\mathbf v \wedge \mathbf w)

Both the cross and wedge products of two identical vectors are zero:

\mathbf u \times \mathbf u = 0
\mathbf u \wedge \mathbf u = 0

\mathbf u \times \mathbf v is perpendicular to the plane containing \mathbf u and \mathbf v.
\mathbf u \wedge \mathbf v is an oriented representation of the same plane.

The cross product of traditional vector algebra (on \mathbb{R}^3) find its place in geometric algebra \mathcal{G}_3 as a scaled exterior product

{a}\times{b} = -i({a}\wedge{b})

(this is antisymmetric). Relevant is the distinction between axial and polar vectors in vector algebra, which is natural in geometric algebra as the distinction between vectors and bivectors (elements of grade two).

The i here is a unit pseudoscalar of Euclidean 3-space, which establishes a duality between the vectors and the bivectors, and is named so because of the expected property

i2 = (e1e2e3)2 = e1e2e3e1e2e3 = − e1e2e1e3e2e3 = e1e1e2e3e2e3 = − e3e2e2e3 = − 1

The equivalence of the \mathbb{R}^3 cross product and the wedge product expression above can be confirmed by direct multiplication of i = − e1e2e3 with a determinant expansion of the wedge product

 u \wedge  v = \sum_{1\leq i<j\leq 3}(u_i v_j - v_i u_j)  {e_i} \wedge {e_j}
= \sum_{1\leq i<j\leq 3}(u_i v_j - v_i u_j)  {e_i}  {e_j}

See also Cross product as an exterior product. Essentially, the geometric product of a bivector and the pseudoscalar of Euclidean 3-space provides a method of calculation of the Hodge dual.

Norm of a vector

The norm (length) of a vector is defined in terms of the dot product

 {\Vert \mathbf u \Vert}^2 = \mathbf u \cdot \mathbf u

Using the geometric product this is also true, but this can be also be expressed more compactly as


{\Vert \mathbf u \Vert}^2 = {\mathbf u}^2

This follows from the definition of the geometric product and the fact that a vector wedge product with itself is zero

 \mathbf u \, \mathbf u = \mathbf u \cdot \mathbf u + \mathbf u \wedge \mathbf u = \mathbf u \cdot \mathbf u

Lagrange identity

In three dimensions the product of two vector lengths can be expressed in terms of the dot and cross products


{\Vert \mathbf{u} \Vert}^2 {\Vert \mathbf{v} \Vert}^2
=
({\mathbf{u} \cdot \mathbf{v}})^2 + {\Vert \mathbf{u} \times \mathbf{v} \Vert}^2

The corresponding generalization expressed using the geometric product is


{\Vert \mathbf{u} \Vert}^2 {\Vert \mathbf{v} \Vert}^2
= ({\mathbf{u} \cdot \mathbf{v}})^2 - (\mathbf{u} \wedge \mathbf{v})^2

This follows from expanding the geometric product of a pair of vectors with its reverse


(\mathbf{u} \mathbf{v})(\mathbf{v} \mathbf{u}) 
= ({\mathbf{u} \cdot \mathbf{v}} + {\mathbf{u} \wedge \mathbf{v}}) ({\mathbf{u} \cdot \mathbf{v}} - {\mathbf{u} \wedge \mathbf{v}})

Determinant expansion of cross and wedge products


\mathbf u \times \mathbf v = \sum_{i<j}{ \begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}  {\mathbf e}_i \times {\mathbf e}_j }

\mathbf u \wedge \mathbf v = \sum_{i<j}{ \begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}  {\mathbf e}_i \wedge {\mathbf e}_j }

Without justification or historical context, traditional linear algebra texts will often define the determinant as the first step of an elaborate sequence of definitions and theorems leading up to the solution of linear systems, Cramer's rule and matrix inversion.

An alternative treatment is to axiomatically introduce the wedge product, and then demonstrate that this can be used directly to solve linear systems. This is shown below, and does not require sophisticated math skills to understand.

It is then possible to define determinants as nothing more than the coefficients of the wedge product in terms of "unit k-vectors" ({\mathbf e}_i \wedge {\mathbf e}_j terms) expansions as above.

A one by one determinant is the coefficient of \mathbf{e}_1 for an \mathbb R^1 1-vector.
A two-by-two determinant is the coefficient of \mathbf{e}_1 \wedge \mathbf{e}_2 for an \mathbb R^2 bivector
A three-by-three determinant is the coefficient of \mathbf{e}_1 \wedge \mathbf{e}_2 \wedge \mathbf{e}_3 for an \mathbb R^3 trivector
...

When linear system solution is introduced via the wedge product, Cramer's rule follows as a side effect, and there is no need to lead up to the end results with definitions of minors, matrices, matrix invertibility, adjoints, cofactors, Laplace expansions, theorems on determinant multiplication and row column exchanges, and so forth.

Matrix Related

Matrix inversion (Cramer's rule) and determinants can be naturally expressed in terms of the wedge product.

The use of the wedge product in the solution of linear equations can be quite useful for various geometric product calculations.

Traditionally, instead of using the wedge product, Cramer's rule is usually presented as a generic algorithm that can be used to solve linear equations of the form Ax = b (or equivalently to invert a matrix). Namely

 x = \frac{1}{| A|}\operatorname{adj}( A) b.

This is a useful theoretic result. For numerical problems row reduction with pivots and other methods are more stable and efficient.

When the wedge product is coupled with the Clifford product and put into a natural geometric context, the fact that the determinants are used in the expression of {\mathbb R}^N parallelogram area and parallelepiped volumes (and higher dimensional generalizations of these) also comes as a nice side effect.

As is also shown below, results such as Cramer's rule also follow directly from the property of the wedge product that it selects non identical elements. The end result is then simple enough that it could be derived easily if required instead of having to remember or look up a rule.

Two variables example


\begin{bmatrix} a &  b\end{bmatrix}
\begin{bmatrix}x \\ y\end{bmatrix}
=  a x +  b y =  c

Pre and post multiplying by a and b

      (  a x +  b y ) \wedge  b = ( a \wedge  b) x =        c \wedge  b
 a \wedge (  a x +  b y )       = ( a \wedge  b) y =  a \wedge  c

Provided  a \wedge  b \neq 0 the solution is

\begin{bmatrix}x \\ y\end{bmatrix}
= \frac{1}{ a \wedge  b}
\begin{bmatrix} c \wedge  b \\  a \wedge  c\end{bmatrix}

For  a,  b \in {\mathbb R}^2, this is Cramer's rule since the {e}_1 \wedge {e}_2 factors of the wedge products

 u \wedge  v = \begin{vmatrix}u_1 & u_2 \\ v_1 & v_2 \end{vmatrix} {e}_1 \wedge {e}_2

divide out.

Similarly, for three, or N variables, the same ideas hold


\begin{bmatrix} a &  b &  c\end{bmatrix}
\begin{bmatrix}x \\ y \\ z\end{bmatrix} =  d

\begin{bmatrix}x \\ y \\ z\end{bmatrix} = \frac{1}{ a \wedge  b \wedge  c}
\begin{bmatrix}
 d \wedge  b \wedge  c \\
 a \wedge  d \wedge  c \\
 a \wedge  b \wedge  d
\end{bmatrix}

Again, for the three variable three equation case this is Cramer's rule since the {e}_1 \wedge {e}_2 \wedge {e}_3 factors of all the wedge products divide out, leaving the familiar determinants.

A numeric example with three equations and two unknowns When there are more equations than variables case, if the equations have a solution, each of the k-vector quotients will be scalars

To illustrate here is the solution of a simple example with three equations and two unknowns.


\begin{bmatrix} 1 \\ 1 \\ 0 \end{bmatrix}
x + 
\begin{bmatrix} 1 \\ 1 \\ 1 \end{bmatrix}
y = 
\begin{bmatrix} 1 \\ 1 \\ 2 \end{bmatrix}

The right wedge product with (1,1,1) solves for x


\begin{bmatrix} 1 \\ 1 \\ 0 \end{bmatrix}
\wedge
\begin{bmatrix} 1 \\ 1 \\ 1 \end{bmatrix}
x = 
\begin{bmatrix} 1 \\ 1 \\ 2 \end{bmatrix}
\wedge
\begin{bmatrix} 1 \\ 1 \\ 1 \end{bmatrix}

and a left wedge product with (1,1,0) solves for y


\begin{bmatrix} 1 \\ 1 \\ 0 \end{bmatrix}
\wedge
\begin{bmatrix} 1 \\ 1 \\ 1 \end{bmatrix}
y = 
\begin{bmatrix} 1 \\ 1 \\ 0 \end{bmatrix}
\wedge
\begin{bmatrix} 1 \\ 1 \\ 2 \end{bmatrix}.

Observe that both of these equations have the same factor, so one can compute this only once (if this was zero it would indicate the system of equations has no solution).

Collection of results for x and y yields a Cramers rule like form:


\begin{bmatrix} x \\ y \end{bmatrix}
=
\frac{1}{(1, 1, 0) \wedge (1, 1, 1)}
\begin{bmatrix}
(1, 1, 2) \wedge (1, 1, 1) \\
(1, 1, 0) \wedge (1, 1, 2)
\end{bmatrix}.

Writing {e} _i \wedge {e} _j = {e} _{ij}, we have the end result:


\begin{bmatrix} x \\ y \end{bmatrix}
=
\frac{1}{{e}_{13} + {e}_{23}}
\begin{bmatrix}
{-{e}_{13} - {e}_{23}} \\
{2{e}_{13} +2{e}_{23}} \\
\end{bmatrix}
=
\begin{bmatrix} -1 \\ 2 \end{bmatrix}.

Equation of a plane

For the plane of all points {\mathbf r} through the plane passing through three independent points {\mathbf r}_0, {\mathbf r}_1, and {\mathbf r}_2, the normal form of the equation is

(({\mathbf r}_2 - {\mathbf r}_0) \times ({\mathbf r}_1 - {\mathbf r}_0)) \cdot ({\mathbf r} - {\mathbf r}_0) = 0.

The equivalent wedge product equation is

({\mathbf r}_2 - {\mathbf r}_0) \wedge ({\mathbf r}_1 - {\mathbf r}_0) \wedge ({\mathbf r} - {\mathbf r}_0) = 0.

Projection and rejection

Using the Gram–Schmidt process a single vector can be decomposed into two components with respect to a reference vector, namely the projection onto a unit vector in a reference direction, and the difference between the vector and that projection.

With,   \hat{u} =  u / {\Vert  u \Vert}, the projection of v onto   \hat{u} is

 \mathrm{Proj}_{{\hat{u}}}\,{v} =  \hat{u} ( \hat{u} \cdot  v)

Orthogonal to that vector is the difference, designated the rejection,

  v -  \hat{u} ( \hat{u} \cdot  v) = \frac{1}{{\Vert  u \Vert}^2} ( {\Vert  u \Vert}^2  v -  u ( u \cdot  v))

The rejection can be expressed as a single geometric algebraic product in a few different ways


 \frac{  u }{{ u}^2} (  u  v -  u \cdot  v)
= \frac{1}{ u} (  u \wedge  v )
=  \hat{u} (  \hat{u} \wedge  v )
= (  v \wedge  \hat{u} )  \hat{u}

The similarity in form between the projection and the rejection is notable. The sum of these recovers the original vector

  v
=  \hat{u} ( \hat{u} \cdot  v) +  \hat{u} (  \hat{u} \wedge  v )

Here the projection is in its customary vector form. An alternate formulation is possible that puts the projection in a form that differs from the usual vector formulation

  v
=  \frac{1}{ u} ( {u} \cdot  v) + \frac{1}{ u} (  {u} \wedge  v )
=  ( {v} \cdot  u) \frac{1}{ u}  + (  v \wedge  u ) \frac{1}{ u}

Working backwards from the end result, it can be observed that this orthogonal decomposition result can in fact follow more directly from the definition of the geometric product itself.


 v =  \hat{u}  \hat{u}  v
=  \hat{u} ( \hat{u} \cdot  v +  \hat{u} \wedge  v )

With this approach, the original geometrical consideration is not necessarily obvious, but it is a much quicker way to get at the same algebraic result.

However, the hint that one can work backwards, coupled with the knowledge that the wedge product can be used to solve sets of linear equations (see: [1] ), the problem of orthogonal decomposition can be posed directly,

Let v = au + x, where  u \cdot  x = 0. To discard the portions of v that are colinear with u, take the wedge product

 u \wedge  v =  u \wedge (a  u +  x) =  u \wedge  x

Here the geometric product can be employed

 u \wedge  v =  u \wedge  x =  u  x -  u \cdot  x =  u  x

Because the geometric product is invertible, this can be solved for x

 x = \frac{1}{ u}( u \wedge  v)

The same techniques can be applied to similar problems, such as calculation of the component of a vector in a plane and perpendicular to the plane.


For three dimensions the projective and rejective components of a vector with respect to an arbitrary non-zero unit vector, can be expressed in terms of the dot and cross product

\mathbf v = (\mathbf v \cdot \hat{\mathbf u})\hat{\mathbf u} + \hat{\mathbf u} \times (\mathbf v \times \hat{\mathbf u})

For the general case the same result can be written in terms of the dot and wedge product and the geometric product of that and the unit vector

\mathbf v = (\mathbf v \cdot \hat{\mathbf u})\hat{\mathbf u} + (\mathbf v \wedge \hat{\mathbf u}) \hat{\mathbf u}

It's also worthwhile to point out that this result can also be expressed using right or left vector division as defined by the geometric product

\mathbf v = (\mathbf v \cdot \mathbf u)\frac{1}{\mathbf u} + (\mathbf v \wedge \mathbf u) \frac{1}{\mathbf u}
\mathbf v = \frac{1}{\mathbf u}(\mathbf u \cdot \mathbf v) + \frac{1}{\mathbf u}(\mathbf u \wedge \mathbf v)

Like vector projection and rejection, higher dimensional analogs of that calculation are also possible using the geometric product.

As an example, one can calculate the component of a vector perpendicular to a plane and the projection of that vector onto the plane.

Let w = au + bv + x, where  u \cdot  x =  v \cdot  x = 0. As above, to discard the portions of w that are colinear with u or v, take the wedge product

 w \wedge  u \wedge  v = (a  u + b  v +  x) \wedge  u \wedge  v =  x \wedge  u \wedge  v.

Having done this calculation with a vector projection, one can guess that this quantity equals  x ( u \wedge  v). One can also guess there is a vector and bivector dot product like quantity such that the allows the calculation of the component of a vector that is in the "direction of a plane". Both of these guesses are correct, and validating these facts is worthwhile. However, skipping ahead slightly, this to-be-proved fact allows for a nice closed form solution of the vector component outside of the plane:

 x
= ( w \wedge  u \wedge  v)\frac{1}{ u \wedge  v}
= \frac{1}{ u \wedge  v}( u \wedge  v  \wedge  w).

Notice the similarities between this planar rejection result a the vector rejection result. To calculation the component of a vector outside of a plane we take the volume spanned by three vectors (trivector) and "divide out" the plane.

Independent of any use of the geometric product it can be shown that this rejection in terms of the standard basis is

 x = \frac{1}{(A_{u,v})^2} \sum_{i<j<k}
\begin{vmatrix}w_i & w_j & w_k \\u_i & u_j & u_k \\v_i & v_j & v_k \\\end{vmatrix}
\begin{vmatrix}u_i & u_j & u_k \\v_i & v_j & v_k \\ { e}_i & { e}_j & { e}_k \\ \end{vmatrix}

where

(A_{u,v})^2
= \sum_{i<j} \begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}
= -( u \wedge  v)^2

is the squared area of the parallelogram formed by u, and v.

The (squared) magnitude of x is

{\Vert  x \Vert}^2 =
 x \cdot  w =
\frac{1}{(A_{u,v})^2} \sum_{i<j<k}
{\begin{vmatrix}w_i & w_j & w_k \\u_i & u_j & u_k \\v_i & v_j & v_k \\\end{vmatrix}}^2

Thus, the (squared) volume of the parallelopiped (base area times perpendicular height) is


\sum_{i<j<k}
{\begin{vmatrix}w_i & w_j & w_k \\u_i & u_j & u_k \\v_i & v_j & v_k \\\end{vmatrix}}^2

Note the similarity in form to the w, u,v trivector itself


\sum_{i<j<k}
{\begin{vmatrix}w_i & w_j & w_k \\u_i & u_j & u_k \\v_i & v_j & v_k \\\end{vmatrix}} { e}_i \wedge { e}_j \wedge { e}_k

which, if you take the set of { e}_i \wedge { e}_j \wedge { e}_k as a basis for the trivector space, suggests this is the natural way to define the length of a trivector. Loosely speaking the length of a vector is a length, length of a bivector is area, and the length of a trivector is volume.

If a vector is factored directly into projective and rejective terms using the geometric product  v = \frac{1}{ u}(  u \cdot  v +  u \wedge  v), then it is not necessarily obvious that the rejection term, a product of vector and bivector is even a vector. Expansion of the vector bivector product in terms of the standard basis vectors has the following form

Let 
 r
= \frac{1}{ u} ( u \wedge  v )
= \frac{ u}{ u^2} (  u \wedge  v ) 
= \frac{1}{{\Vert  u \Vert}^2}  u (  u \wedge  v )

It can be shown that


 r = \frac{1}{{\Vert{ u}\Vert}^2} \sum_{i<j}\begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}
\begin{vmatrix}u_i & u_j\\  e_i &  e_j\end{vmatrix}

(a result that can be shown more easily straight from  r =  v -  \hat{u} ( \hat{u} \cdot  v)).

The rejective term is perpendicular to u, since \begin{vmatrix}u_i & u_j\\ u_i & u_j\end{vmatrix} = 0 implies \ r \cdot  u =  0 .

The magnitude of r, is


{\Vert  r \Vert}^2 =  r \cdot  v = \frac{1}{{\Vert{ u}\Vert}^2} \sum_{i<j}\begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}^2
.

So, the quantity


{\Vert  r \Vert}^2 {\Vert{ u}\Vert}^2 = \sum_{i<j}\begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}^2

is the squared area of the parallelogram formed by u and v.

It is also noteworthy that the bivector can be expressed as


 u \wedge  v = \sum_{i<j}{ \begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}   e_i \wedge  e_j }
.

Thus is it natural, if one considers each term  e_i \wedge  e_j as a basis vector of the bivector space, to define the (squared) "length" of that bivector as the (squared) area.

Going back to the geometric product expression for the length of the rejection \frac{1}{ u} (  u \wedge  v ) we see that the length of the quotient, a vector, is in this case is the "length" of the bivector divided by the length of the divisor.

This may not be a general result for the length of the product of two k-vectors, however it is a result that may help build some intuition about the significance of the algebraic operations. Namely,

When a vector is divided out of the plane (parallelogram span) formed from it and another vector, what remains is the perpendicular component of the remaining vector, and its length is the planar area divided by the length of the vector that was divided out.

Area of the parallelogram defined by u and v

If A is the area of the parallelogram defined by u and v, then


A^2 = {\Vert \mathbf u \times \mathbf v \Vert}^2 = \sum_{i<j}{\begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}}^2,

and


A^2 = -(\mathbf u \wedge \mathbf v)^2 = \sum_{i<j}{\begin{vmatrix}u_i & u_j\\v_i & v_j\end{vmatrix}}^2.

Note that this squared bivector is a geometric multiplication; this computation can alternatively be stated as the Gram determinant of the two vectors.

Angle between two vectors

({\sin \theta})^2 = \frac{{\Vert \mathbf u \times \mathbf v \Vert}^2}{{\Vert \mathbf u \Vert}^2 {\Vert \mathbf v \Vert}^2}
({\sin \theta})^2 = -\frac{(\mathbf u \wedge \mathbf v)^2}{{ \mathbf u }^2 { \mathbf v }^2}

Volume of the parallelopiped formed by three vectors

In vector algebra, the volume of a parallelopiped is given by the square root of the squared norm of the scalar triple product:

V^2 = {\Vert (\mathbf u \times \mathbf v) \cdot \mathbf w \Vert}^2
= {\begin{vmatrix}
u_1 & u_2 & u_3 \\ 
v_1 & v_2 & v_3 \\ 
w_1 & w_2 & w_3 \\ 
\end{vmatrix}}^2

V^2 = -(\mathbf u \wedge \mathbf v \wedge \mathbf w)^2
= -\left(\sum_{i<j<k}
\begin{vmatrix}
u_i & u_j & u_k \\ 
v_i & v_j & v_k \\ 
w_i & w_j & w_k \\ 
\end{vmatrix}
\hat{\mathbf e}_i \wedge
\hat{\mathbf e}_j \wedge
\hat{\mathbf e}_k 
\right)^2

= \sum_{i<j<k}
{\begin{vmatrix}
u_i & u_j & u_k \\ 
v_i & v_j & v_k \\ 
w_i & w_j & w_k \\ 
\end{vmatrix}}^2

Product of a vector and a bivector

In order to justify the normal to a plane result above, a general examination of the product of a vector and bivector is required. Namely,

 w ( u \wedge  v)
= \sum_{i,j<k}w_i { e}_i {\begin{vmatrix}u_j & u_k \\v_j & v_k \\\end{vmatrix}} { e}_j \wedge { e}_k

This has two parts, the vector part where i = j or i = k, and the trivector parts where no indexes equal. After some index summation trickery, and grouping terms and so forth, this is

 w ( u \wedge  v) = 
\sum_{i<j}(w_i { e}_j 
- w_j { e}_i )
{\begin{vmatrix}u_i & u_j \\v_i & v_j \\\end{vmatrix}}

+
\sum_{i<j<k}
{\begin{vmatrix}w_i & w_j & w_k \\ u_i & u_j & u_k \\v_i & v_j & v_k \\\end{vmatrix}} 
{ e}_i \wedge { e}_j \wedge { e}_k

The trivector term is  w \wedge  u \wedge  v. Expansion of ( u \wedge  v)  w yields the same trivector term (it is the completely symmetric part), and the vector term is negated. Like the geometric product of two vectors, this geometric product can be grouped into symmetric and antisymmetric parts, one of which is a pure k-vector. In analogy the antisymmetric part of this product can be called a generalized dot product, and is roughly speaking the dot product of a "plane" (bivector), and a vector.

The properties of this generalized dot product remain to be explored, but first here is a summary of the notation

 w ( u \wedge  v) =  w \cdot ( u \wedge  v) +  w \wedge  u \wedge  v
( u \wedge  v)  w = -  w \cdot ( u \wedge  v) +  w \wedge  u \wedge  v
 w \wedge  u \wedge  v = \frac{1}{2}( w ( u \wedge  v) + ( u \wedge  v)  w)
 w \cdot ( u \wedge  v) = \frac{1}{2}( w ( u \wedge  v) - ( u \wedge  v)  w)

Let w = x + y, where x = au + bv, and  y \cdot  u =  y \cdot  v =  0. Expressing w and the  u \wedge  v, products in terms of these components is


 w ( u \wedge  v) =  x ( u \wedge  v) +  y ( u \wedge  v)
= 
 x \cdot ( u \wedge  v) +  y \cdot ( u \wedge  v) +  y \wedge  u \wedge  v

With the conditions and definitions above, and some manipulation, it can be shown that the term  y \cdot ( u \wedge  v) =  0, which then justifies the previous solution of the normal to a plane problem. Since the vector term of the vector bivector product the name dot product is zero when the vector is perpendicular to the plane (bivector), and this vector, bivector "dot product" selects only the components that are in the plane, so in analogy to the vector-vector dot product this name itself is justified by more than the fact this is the non-wedge product term of the geometric vector-bivector product.

Derivative of a unit vector

It can be shown that a unit vector derivative can be expressed using the cross product


\frac{d}{dt}\left(\frac{\mathbf r}{\Vert \mathbf r \Vert}\right)
= \frac{1}{{\Vert \mathbf r \Vert}^3}\left(\mathbf r \times \frac{d \mathbf r}{dt}\right) \times \mathbf r
= \left(\hat{\mathbf r} \times \frac{1}{{\Vert \mathbf r \Vert}} \frac{d \mathbf r}{dt}\right) \times \hat{\mathbf r}

The equivalent geometric product generalization is


\frac{d}{dt}\left(\frac{\mathbf r}{\Vert \mathbf r \Vert}\right)
= \frac{1}{{\Vert \mathbf r \Vert}^3}\mathbf r \left(\mathbf r \wedge \frac{d \mathbf r}{dt}\right)
= \frac{1}{{ \mathbf r }}\left(\hat{\mathbf r} \wedge \frac{d \mathbf r}{dt}\right)

Thus this derivative is the component of \frac{1}{{\Vert \mathbf r \Vert}}\frac{d \mathbf r}{dt} in the direction perpendicular to \mathbf r. In other words this is \frac{1}{{\Vert \mathbf r \Vert}}\frac{d \mathbf r}{dt} minus the projection of that vector onto \hat{\mathbf r}.

This intuitively makes sense (but a picture would help) since a unit vector is constrained to circular motion, and any change to a unit vector due to a change in its generating vector has to be in the direction of the rejection of \hat{\mathbf r} from \frac{d \mathbf r}{dt}. That rejection has to be scaled by 1/|r| to get the final result.

When the objective isn't comparing to the cross product, it's also notable that this unit vector derivative can be written


{{ \mathbf r }} \frac{d \hat{\mathbf r}}{dt}
= \hat{\mathbf r} \wedge \frac{d \mathbf r}{dt}

See also


Wikimedia Foundation. 2010.

Игры ⚽ Поможем написать реферат

Look at other dictionaries:

  • Geometric algebra — In mathematical physics, a geometric algebra is a multilinear algebra described technically as a Clifford algebra over a real vector space equipped with a non degenerate quadratic form. Informally, a geometric algebra is a Clifford algebra that… …   Wikipedia

  • algebra — /al jeuh breuh/, n. 1. the branch of mathematics that deals with general statements of relations, utilizing letters and other symbols to represent specific sets of numbers, values, vectors, etc., in the description of such relations. 2. any of… …   Universalium

  • Eigenvalue, eigenvector and eigenspace — In mathematics, given a linear transformation, an Audio|De eigenvector.ogg|eigenvector of that linear transformation is a nonzero vector which, when that transformation is applied to it, changes in length, but not direction. For each eigenvector… …   Wikipedia

  • List of terms relating to algorithms and data structures — The [http://www.nist.gov/dads/ NIST Dictionary of Algorithms and Data Structures] is a reference work maintained by the U.S. National Institute of Standards and Technology. It defines a large number of terms relating to algorithms and data… …   Wikipedia

  • analysis — /euh nal euh sis/, n., pl. analyses / seez /. 1. the separating of any material or abstract entity into its constituent elements (opposed to synthesis). 2. this process as a method of studying the nature of something or of determining its… …   Universalium

  • mathematics — /math euh mat iks/, n. 1. (used with a sing. v.) the systematic treatment of magnitude, relationships between figures and forms, and relations between quantities expressed symbolically. 2. (used with a sing. or pl. v.) mathematical procedures,… …   Universalium

  • Quaternion — Quaternions, in mathematics, are a non commutative extension of complex numbers. They were first described by the Irish mathematician Sir William Rowan Hamilton in 1843 and applied to mechanics in three dimensional space. They find uses in both… …   Wikipedia

  • Classical Hamiltonian quaternions — For the history of quaternions see:history of quaternions For a more general treatment of quaternions see:quaternions William Rowan Hamilton invented quaternions, a mathematical entity in 1843. This article describes Hamilton s original treatment …   Wikipedia

  • Maxwell's equations — For thermodynamic relations, see Maxwell relations. Electromagnetism …   Wikipedia

  • List of mathematics articles (C) — NOTOC C C closed subgroup C minimal theory C normal subgroup C number C semiring C space C symmetry C* algebra C0 semigroup CA group Cabal (set theory) Cabibbo Kobayashi Maskawa matrix Cabinet projection Cable knot Cabri Geometry Cabtaxi number… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”