Ortho Spaces
Ortho Spaces
ORTHOGONAL GROUPS
An Introduction to the Concepts Required in the
Theory of the Representation of the Euclidean
Group.
José Marı́a Rico
Instituto Tecnológico de Celaya,
Departamento de Ingenierı́a Mecánica
Celaya, Gto. 38000, México
May 14, 2018
1 Preface
The purpose of these notes is to gather in a single work concepts and results
that play an important role in the theory of representations of the Euclidean
group, and hence in the kinematics of the rigid body. Most of the results
presented here can be found scattered in the specialized literature. However, as
far as the author is aware, Theorem 7 in Section 4 is a new generalization of a
result by Porteous. Furthermore, most of Section 7, on orthogonal mappings of
degenerate spaces, is thought to be new.
1
(a, µ1 b1 + µ2 b2 ) = µ1 (a, b1 ) + µ2 (a, b2 ).
Since the vector space is real, and the mapping is upon the real field, the sym-
metric bilinear form is called real. Only real symmetric bilinear forms will be
considered here.
Definition 2. Cuadratic Form. The quadratic form associated with a
symmetric bilinear form is the map
X −→ R a −→ a · a = (a, a).
This definition implies that a symmetric bilinear form uniquely defines a quadratic
form. The following results simply shows that the opposite is also true.
Theorem 3. A symmetric bilinear form is uniquely defined by its quadratic
form.
Proof. Let (a − b) ∈ X; then
Therefore
1
(a, b) =
[(a, a) + (b, b) − (a − b, a − b)]
2
Corollary 4. The symmetric bilinear form is the zero map if, and only if,
its quadratic form is the zero map.
Definition 5. Two elements a, b ∈ X are said to be mutually orthogonal if
(a, b) = a · b = 0.
(a, b) = a · b = 0 ∀a ∈ A, b ∈ B.
2
Finally, an orthogonal space is said to be indefinite if there exists a, b ∈ X such
that
(a, a) > 0 and (b, b) < 0.
An orthogonal space whose symmetric scalar product (symmetric bilinear form)
is the zero map is said to be isotropic, and an orthogonal space that is the direct
sum of two isotropic subspaces is said to be neutral.
Examples.
1. The vector space Rp+q together with the symmetric bilinear form
p
X q
X
(a, b) = − a i bi + ap+j bp+j .
i=1 j=1
2. The vector space R2n together with the symmetric bilinear form
n
X
(a, b) = (ai bn+i + an+i bi )
i=1
3
Proof. Firstly consider
σ : X −→ X L σ(x) = xσ
or
xσ ( ) = 2x1 ( )1 − 3x2 ( )3 + 5x3 ( )2 .
This is a linear correlation. However it is not symmetric.2
Theorem 2. Any linear symmetric correlation σ induces a symmetric bi-
linear form in the following way
4
Conversely, any scalar product (a, b) induces a unique symmetric correlation,
namely the map a −→ aσ , where the functional aσ is defined by
aσ (b) = (a, b) ∀b ∈ X.
Assume that (a, b) is a symmetric bilinear form in X, and define the correlation
σ : X −→ X L as σ(a) = aσ where aσ is defined as
5
Definition 3. Let X be a finite-dimensional real orthogonal space with
induced correlation σ. Since σ : X −→ X L , and dimX = dimX L , then σ is
injective if, and only if, σ is bijective. If σ is bijective, then the orthogonal
space, its symmetric bilinear form, its quadratic form, and its correlation are
said to be non-degenerate or non-singular. If, on the other hand, σ is not
bijective, then σ is not injective, and ker σ 6= {0}. Then X, the correlation σ
and the induced symmetric bilinear form are said to be degenerate, and the
kernel of σ is also called the kernel4 of X.
Theorem 4. The kernel of X is a subspace, and it contains each and every
one of the elements of X which are orthogonal to X itself.
Proof. First, it will be enlightening to characterize explicitly the elements
of the kernel of X. An a ∈ X belongs to the kernel of σ if, and only if, σ(a) = aσ
is the zero functional, this is to say aσ (b) = 0 for every b ∈ X. Consider now
a, b ∈ kerX; then
σ(a + b) = σ(a) + σ(b) = 0 + 0 = 0
Hence a + b ∈ kerX. Similary, if µ ∈ R, then
σ(µa) = µσ(a) = µ0 = 0.
aσ (x) = (a, x) = 0 ∀x ∈ X;
or
xσ ( ) = 2x1 ( )1 − 3x3 ( )3
3
Then x ∈ kerℜ if, and only if, x is the zero map; i.e. xσ (y) = 0 for every
σ
xσ (ei ) = 0 f or i = 1, 2, 3.
Therefore
6
Thus x ∈ kerℜ3 if, and only if, x1 = x3 = 0. Hence
Consider then
and
(ei , ei ) = +1 for every 1 = p + a, ..., p + q
Hence the e′i sare invertible and the conditions of the previous theorem apply.
Theorem 8. Let X be a finite-dimensional real orthogonal space, and let
X ′ be a linear complement in X of kerX. Then X ′ is a non-degenarate subspace
of X.
Proof. Assume by the way of contradiction, that there exists a non-zero
x ∈ X ′ such that
(x, y) = 0 ∀y ∈ X ′
Since, by definition of kerX, (x, z) = 0 for every z ∈ kerX, and X = kerX ⊕ X ′
(this assumption implies that X ′ ∩ kerX = {0} then
(x, w) = 0 ∀w ∈ X.
7
4 Orthogonal Maps.
Definition 1. Let X an Y be real orthogonal spaces with induced correlations
σ and τ respectively. A map t : X → Y is said to be an orthogonal map if it is
linear and, for all a, b ∈ X
However
8
Proof. The linearity of these mappings is easily proved. Here, only the
preservation of the symmetric bilinear form will be proved. Let a, b ∈ X be
arbitrary; the IX (a) = a, and IX (b) = b thus
Thus
(y1 , y2 ) = (x1 , x2 ) = (t−1 (y1 ), t−1 y2 ),
and t−1 is orthogonal.
Definition 5. An invertible orthogonal map t : X → Y will be called an
orthogonal isomorphism, and the two orthogonal space X and Y are said to be
isomorphic.
2n
Theorem 6. For any finite n, the orthogonal spaces Rn,n and Rhb are
isomorphic.
Definition 6. An invertible orthogonal map t : X → X is called an or-
thogonal automorphism of the orthogonal space. It is important to recall
that since both, the range and the image of the mapping t are the same, the
invertibility of the mapping is assured by the injectivity or the surjectivity of
the mapping.
Theorem 8. Any orthogonal mapping of a non-degenerate finite-dimensional
orthogonal space X into itself is an orthogonal automorphism of X.
Proof. By Theorem 2, any orthogonal mapping of a non-degenerate finite-
dimensional orthogonal space X is injective; therefore the mapping is surjective
and bijective.
Definition 9. Let X and Y be orthogonal spaces; the set of orthogonal
maps t : X → Y will be denoted by O(X, Y ). The set of the orthogonal
automorphisms t : X → X will be denoted by O(X).
Theorem 10. O(X) forms a group under the operation of composition.
This group is called the group of orthogonal automorphisms of X.
Proof. It is well known that mappings, in general, form a group under the
operation of composition. Therefore, only the closure and the presence of the
indentity will be indicated.
9
and
(t−1 s−1 )st = t−1 (s−1 s)t = t−1 IX t = IX
Thus st ∈ O(X).
5 Orthogonal Annihilators
Definition 1. Let W be a subspace of a vector space X, the dual annihilator
of W , denoted W @ , is the subspace of the dual subspace X L that annihilates
W ; i.e.
W @ = {β ∈ X L | β(w) = 0 ∀w ∈ W }
Theorem 2. Let W be a subspace of a vector space X; then
10
Proof. First it is necessary to recall that dimX = dimX L , and arbitrary
element of the dual space is of the form
W ⊥ = σ ⊣ (W @ ) = {x ∈ X | σ(X) ∈ W @ }
aσ (b) = (a, b) = 0 ∀b ∈ W
Thus a ∈ W ⊥ .
Definition 5. Let X be a real orhogonal space and W a subspace of X with
linear complement Y ; i.e. X = W ⊕ Y ; if Y is contained in W ⊥ then we said
that Y is an orthogonal complement of W in X. A direct sum decomposition
X = W ⊕ Y is said to be an orthogonal decomposition.
Example. Consider the orthogonal space R4 with the symmetric bilinear
form
(x, y) = 0x1 y1 + x2 y2 + x3 y3 + x4 y4
Consider W = [(1, 0, 0, 0)]. Then, the orthogonal annihilator of W is W ⊥ = R4 ,
and Y = {(0, y2 , y3 , y4 ) | y2 , y3 , y4 ∈ ℜ} is an orthogonal complement of W in
R4 .
5 Furthermore it can be proved that the equations are independent.
11
Theorem 6. Let W be a linear subspace of a finite-dimensional real or-
thogonal space X. Then kerW = W ∩ W ⊥ .
Proof. By definition
then
kerW = {w | w ∈ W } ∩ {w | (w, w′ ) = 0 ∀w′ ∈ W }.
Thus
kerW = W ∩ W ⊥ .
Theorem 7. Let W be a linear subspace of a finite-dimensional real or-
thogonal space X. Then X = W ⊕ W ⊥ if, and only if, W is non-degenerate; in
this case, W ⊥ is the unique orthogonal complement of W in X.
Proof. Assume W is degenerate then W ∩ W ⊥ 6= {0}, and the sum is not
direct. Assume W is non-degenerate. Then W ∩ W ⊥ = {0}; thus the sum
W + W ⊥ es direct. Then
since
dimW ⊥ = dim[σ ⊣ (W @ )] ≥ dimW @ .
where σ is the induced correlation. Thus
Then
dim(W + W ⊥ ) ≤ dimX.
Therefore
W + W ⊥ = X.
The uniqueness of the orhogonal complement is proved by contradiction. As-
sume that U1 , and U2 are two orthogonal complements, then
X = W ⊕ U1 with U1 ⊂ W ⊥
and
X = W ⊕ U2 with U2 ⊂ W ⊥ .
Let {w1 , w2 , . . . , wk }, {u1 , u2 , . . . , un−k }, and {u′1 , u′2 , . . . , u′n−k } be bases of
W, U1 , and U2 respectively. Since, by assumption U1 6= U2 , there is at least
12
one u′i , let us say u′1 , such that u′1 6∈ U1 . Since X = W + U1 , then there exist
w ∈ W and u ∈ U1 , such that
(w, w′ ) = 0 ∀w′ ∈ W ⊥ .
Thus (W ⊥ )⊥ .
Corollary 10. Let W be a non-degenerate orthogonal subspace of a non-
degenerate orthogonal space X, then W ⊥ is also a non-degenerate orthogonal
subspace of X.
Proof. Since W is non-degenerate, then
W ∩ W ⊥ = {0},
and by theorem 9,
(W ⊥ )⊥ = W.
Thus
W ⊥ ∩ (W ⊥ )⊥ = {0},
and W ⊥ is non-degenerate.
13
6 Basis of Orthogonal Spaces.
Theorem 1. Basis Theorem for Real Orthogonal Spaces. An non-
dimensional non-degenerate real orthogonal space X, with n > 0, is express-
ible as the direct sum of n non-degenerate mutually orthogonal lines (one-
dimensional subspaces of X).
Proof. By induction, if n = 1, the space consists of only one line, and by
assumption the line is non-degenerate. Assume that the theorem is true for n
and consider the case when the dimension is n + 1.
Since the space is non-degenerate, by the theorem 2.9, there exists and in-
vertible element a ∈ X that defines a non-degenerate line R[a], such that
X = (R[a]) ⊕ (R[a])⊥
By the corollary 5.10 (R[a])⊥ is also non-degenarate and of dimension n, thus
the proof is finished.
Definition 2. A linearly independent subset S of a real orthogonal space
X is said to be an Orthonormal Subset of X if two distinct elements of S
are mutually orthogonal, and the quadratic form of any element of S is equal
to −1, 0, or 1. If S also spans X, then S is said to be an Orthonormal Basis
of X.
Theorem 3. Basis Theorem. Any finite-dimensional orthogonal space X
has an orthonormal basis.
Proof. Let X ′ be a linear complement, in X, of kerX. Then X ′ is a non-
degenarate subspace of X, and, by theorem 1, it has a basis B ′ such that is
orthogonal and for any b′i ∈ B ′ satisfies (b′i , b′i ) 6= 0. Then, by applying theorem
2.10 it is possible to find a basis B, of X ′ , orthogonal and with
(bi , bi ) ∈ (−1, 1)
Let A be any basis for kerX, then A ∪ B is an orthogonal basis, such that their
elements ci satisfy
(ci , ci ) ∈ (−1, 0, 1)
Thus A ∪ B is an orthonormal basis of X.
Corollary 4. Any non-degenerate finite-dimensional orthogonal space X is
isomorphic to Rp,q for some finite p, and q.
Corollary 5. Any orthogonal finite-dimensional space X is isomorphic to
Rp,q,r , where the symmetric bilinear form is
(Rp,q,r , Rp,q,r ) = R
p
X q
X r
X
(x, y) = x i yi − xp+j yp+j + xp+q+k yp+q+k
i=1 j=1 k=1
Furthermore, the orthogonal space X can be orthogonally decomposed as
X = kerX ⊕ W1 ⊕ W2
where dim(kerX) = p, dimW1 = q, and dimW2 = r. Further W1 is negative
definite, and W2 is positive definite.
14
7 Reflections and Orthogonal Mappings
Theorem 1. Let W ⊕ Y be an orthogonal decomposition of an orthogonal
space X. Then the mapping
σ : X → X; x=w+y w−y
v 1 = w1 + y1 , and v 2 = w2 + y2 .
15
Theorem 5. Suppose that a and b are invertible elements of a finite-
dimensional orthogonal space X, such that (a, a) = (b, b). Thena may be
mapped into b either by a single hyperplane reflections of X.
Proof. By theorem 2.11, either a − b or a + b are invertible. Furthermore
a − b and a + b are orthogonal since
t = u−1 (ut)R ,
16
8 Orthogonal Mappings of Degenerate Spaces
Corollary 5, section 6, establishes that any finite-dimensional orthogonal space
X can be orthogonally decomposed as
X = kerX ⊕ W1 ⊕ W2
The first step in the analysis of orthogonal mappings of degenerate spaces consist
in analyzing the effect of the mapping into these subspaces.
Theorem 1. Let t : X −→ X be an orthogonal mapping of a degener-
ate orthogonal space, then kerX, W1 , and W2 are invariant subspaces of t.
Furthermore, t(W1 ) = W1 , and t(W2 ) = W2 .
Proof. Let {e1 , ..., ep , ep+1 , . . . , ep+q , ep+q+1 , . . . , en } be an orthonormal ba-
sis of X, with the first step p elements as a basis of kerX, the next q elements
as a basis of W1 , and the remaining ones as a basis of W2 . Let x ∈ W1 ; then
q
X
x= µj ep+j
j=1
t[µ1 ep+1 + . . . + µq ep + q] = 0
Hence,
t(x) = y1 + y2 + y3
17
Proof. Assume that t : X → X is an orthofonal map of the degenerate
orthogonal space. Since by theorem 1, W1 and W2 are invariant subspaces,
then W = W1 ⊕ W2 is also invariant, and it is possible to define the restriction
mapping tW : W → W . Furthermore, for every x, y ∈ W
Furthermore
18
9 Semiorientations and Orthogonal Groups that
Preserve the Semiorientations
Definition 1. Let X be a non-degenerate space; then X is isomorphic to
Rp,q for some integers p, and q. Furthermore, X = W1 ⊕ W2 , where W1 is
negative definite and W2 is positive definite. The semiorientations of X are the
orientations of the subspaces W1 and W2 .
In general, although the elements of SO(X) preserve the orientation of X,
they do not necesarily preserve the orientation of the subspaces W1 and W2
defined in the statement above. The situation is illustrated below.
Example. Consider The orthogonal space R1,2 wich has the symmetric bi-
linear form
tW 1 (x) = t(x) ∀ x ∈ W1
and
tW 2 (x) = t(x) ∀ x ∈ W2
then t preserves the semiorientations of W .
19
Theorem 3. The set of orthogonal automorphisms that preserves the
semiorientatons of a non-degenerate orthogonal space X is a subgroup of SO(X).
This subgroup is denoted by SO+ (X). In particular if X = Rp,q , then SO+ (X)
is denoted by SO+ (p, q). 6
the matrix is symmetric. The other direction of th proof is easy but laborious.
The next proposition shows how the matrix representative of the symmetric
bilinear form can be used to find the value, of the form, for arbitrary vectors of
the space.
Proposition 2. Let X be an orthogonal space with S as the matrix
representative of the symmetric bilinear form with respect to a basis B =
6 The concept of semiorientations can be applied to degenerate orthogonal spaces; however
20
{x1 , x2 , . . . , xn }. Let A = [a1 , a2 , . . . , an ], and C = [c1 , c2 , . . . , cn ] be the co-
ordinate vectors, with respect to the basis B, of two vectors, a, c ∈ X. Then
c1
c2
T
.
(a, c) = [a1 , a2 , . . . , an ] S
.
(3)
.
cn
sij = (xi , xj ) ∀ i, j = 1, 2, . . . , n.
Since
a = a1 x 1 + a2 x 2 + . . . + an x n and c = c1 x1 + c2 x2 + . . . + cn xn ,
then
(a, c) = (a1 x1 + a2 x2 + . . . + an xn , c1 x1 + c2 x2 + . . . + cn xn )
= a1 [c1 (x1 , x1 ) + c2 (x1 , x2 ) + . . . + cn (x1 , xn )] +
a2 [c1 (x2 , x1 ) + c2 (x2 , x2 ) + . . . + cn (x2 , xn )] + . . . +
an [c1 (xn , x1 ) + c2 (xn , x2 ) + . . . + cn (xn , xn )]
= a1 (c1 s11 + c2 s12 + . . . + cn s1n ) +
a2 (c1 s21 + c2 s22 + . . . + cn s2n ) + . . . +
an (c1 sn1 + c2 sn2 + . . . + cn snn )
or
s11 s12 . . . s1n c1
s21 s22 . . . s2n
c2
. . . . . . .
(a, c) = [a1 , a2 , ..., an ]T = AT SC
. . . . . .
.
. . . . . . .
sn1 sn2 . . . snn cn
One important feature of the use of the matrix representative of the symmetric
bilinear form is that allows one to work with more concrete beings than map-
pings; i.e. matrices and coordinate vectors. However, one drawback is that there
is a multitude of matrix representatives, indeed one for each posible basis of the
orthogonal space. Thus a property of the symmetric bilinear form will be trans-
lated into a common property of all the possible representations. Conversely,
only those properties that are common to all possible matrix representations
are properties of the form.
21
The next proposition shows the relationship between the different matrix
representatives of the same symmetric bilinear form.
Proposition 3. Let X be an orthogonal space, and let B = {x1 , x2 , . . . , xn },
′ ′ ′ ′
and B = {x1 , x2 , . . . , xn } be two bases of the space. If S is the matrix repre-
sentative of the symmetric bilinear form with respect to the basis B, then the
matrix representative of the symmetric bilinear form with respect to the basis
′ ′
B , denoted by S , is given by
′
S = P SP T ,
′
where P is the transition matrix between the bases B and B .
′
Proof. Since P is the transition matrix from B to B , then
′
x1 = p11 x1 + p12 x2 + . . . + p1n xn
′
x2 = p21 x1 + p22 x2 + . . . + p2n xn
.
.
.
′
xn = pn1 x1 + pn2 x2 + . . . + pnn xn
′ ′
An arbitrary element sij of S is given by
n n n n
′ ′ ′ X X X X
sij = (xi , xj ) = ( pir xr , pjt xt ) = pir (xr , pjt xt )
r=1 t=1 r=1 t=1
n n n n n
" n
#
X X X X X X
= pir pjt (xr , xt ) = pir pjt srt = pir srt pjt
r=1 t=1 r=1 t=1 r=1 t=1
Pn
Notice that t=1 srt pjt is just the (r − th, j − th) element of the product of
matrices M = SP T , denoted mrj . Then,
n
′ X
sij = pir mrj
r=1
′
Again, sij is the (i − th, j − th) element of the product of the matrices P and
M . Hence ′
S = P M = P SP T .
This result provides the foundation for finding what characteristics are common
to all the representations of a symmetric quadratic form, and therefore these
characteristics are properties of the symmetric bilinear form. An important
example of this class of results is now developed.
′
Proposition 4. Let S ans S be two matrices representative pf a symmetric
′
bilinear form, then S is non-singular (singular) if, and only if, S is non-singular
(singular).
22
Proof. By proposition 3, there exists a non-singular transition matrix P
such that ′
S = P SP T .
Then ′
| S |=| P SP T |=| P || S || P T |
But | P |=| P T |, hence
′
| S |=| S || P |2
′ ′
Since P is non-singular | P |6= 0. Hence | S |6= 0 (| S |= 0) if, and only if,
| S |6= 0 (| S |= 0).
This proposition has shown that the non-singularity (or yhe singularity) of
the matrices representing a symmetric bilinear form is indeed a property of the
symmetric bilinear form. This property calls for a definition a classification of
the symmetric bilinear forms and the orthogonal spaces.
Definition 5. If the matrix representative of a symmetric bilinear form is
non-singular, then the symmetric bilinear form and the orthogonal space are
called non-degenerate. Otherwise, they are called degenerate.
The final step is to reconcile this definition of non-degeneracy, based on the
matrix representatives of the form, with that already given in the main notes;
i.e. it is required to show that two definitions are equivalent.
Proposition 6. The definition of non-degeneracy given here and that given
in the main notes are equivalent.
Proof. According with the definition in the main notes, a symmetric bilinear
form is degenerate if there is a non-zero w ∈ X such that wσ is the zero mapping;
this is to say
wσ (y) ≡ (w, y) = 0 ∀ y ∈ X.
This by proposition 2,
(w, y) = (y, w) = Y T SW = 0 ∀ Y ∈ Rn ,
where S is the matrix representative of the symmetric bilinear form with respect
to a basis B, and W , and Y are the coordinate vectors of w, and y with respect
to the basis B. Therefore
SW = 0
with W 6= 0, and S is singular. Conversely, if S is singular, there exists a W 6= 0
such that
SW = 0;
hence
Y T SW = 0 ∀ Y ∈ Rn .
If w is the vector whose coordinate vector with respect to the basis B is W ,
then
(y, w) = Y T SW = 0 ∀ y ∈ X,
and w ∈ kerσ. Hence, the space is degenerate.
23
10.2 Orthogonal Maps and the Matrix Representation of
the Symmetric Bilinear Form
This section shows how the matrix representation of the symmetric bilinear form
interacts with orthogonal maps; in order to maintain the development within
reasonable limits, it is confined to orthogonal maps of an orthogonal map into
itself.
It is convenient to recall that t : X −→ X is an orthogonal map, of the
orthogonal space X, if
T T ST = S.
24
Let S be the matrix representative of the symmetric bilinear form, with respect
to a basis B. Let A and C be the coordinate vectors of a, and c with respect to
B, and T the matrix representative of t : X −→ X with respect to B. Then
and
(a, c) = AT SC.
Hence, the definition of an orthogonal map leads to
AT T T ST C = AT SC ∀ A, C ∈ Rn ,
and finally
T T ST = S,
and thus t is orthogonal according to the definition given in these notes. The
proof for the other direction can be worked out following the same lines.
References
[1] Kaplansky, I., (1969), Linear Algebra and Geometry a Second Course., 2nd
ed., New York: Chelsea Publishing Company.
[2] Porteus, I. R., (1981), Topological Geometry, 2nd ed., Cambridge: Cam-
bridge University Press.
25