Complexification, complex structures, and linear ordinary differential equations

Jordan Bell
April 3, 2014

1 Motivation

The solution of the initial value problem

x(t)=Ax(t),x(0)=x0n,

where A is an n×n matrix over , is x(t)=exp(At)x0. If we want to compute the solution and if A is diagonalizable, say A=PDP-1, we use

exp(At)=exp((PDP-1)t)=Pexp(Dt)P-1.

Thus if the matrix A has complex eigenvalues, then although exp(At)x0n, it may not be the case that P-1x0n. For example, if A=(0-110), then

D=(-i00i),P=(-ii11),P-1=12(i1-i1).

For x0=(10),

P-1x0=12(i1-i1)(10)=12(i-i).

This is similar to how Cardano’s formula, which expresses the roots of a real cubic polynomial in terms of its coefficients, involves complex numbers and yet the final result may still be real.

In the following, unless I specify the dimension of a vector space, any statement about real vector spaces is about real vector spaces of finite or infinite dimension, and any statement about complex vector spaces is about complex vector spaces of finite or infinite dimension.

2 Direct sums

If V is a real vector space, a complex structure for V is an -linear map J:VV such that J2=-idV.

If V is a real vector space and J:VV is a complex structure, define a complex vector space VJ in the following way: let the set of elements of VJ be V, let addition in VJ be addition in V, and define scalar multiplication in VJ by

(a+ib)v=av+bJ(v).

One checks that for α,β and vVJ we have (αβ)v=α(βv), and thus that VJ is indeed a complex vector space with this definition of scalar multiplication.11 1 One should also verify that distributivity holds with this definition of scalar product; the other properties of a vector space are satisfied because VJ has the same addition as the real vector space V.

Let V be a real vector space, and define the -linear map J:VVVV by

J(v,w)=(-w,v).

J2=-idVV. J is a complex structure on the real vector space VV. The complexification of V is the complex vector space V=(VV)J. Thus, V has the same set of elements as VV, the same addition as VV, and scalar multiplication

(a+ib)(v,w)=a(v,w)+bJ(v,w),

which gives

(a+ib)(v,w)=a(v,w)+b(-w,v)=(av,aw)+(-bw,bv)=(av-bw,aw+bv).

If the real vector space V has dimension n and if {e1,,en} is a basis for V, then

{(e1,0),,(en,0),(0,e1),,(0,en)}

is a basis for the real vector space VV. Let vV. Using the basis for the real vector space VV, there exist

a1,,an,b1,,bn

such that

v = a1(e1,0)+an(en,0)+b1(0,e1)++bn(0,en)
= a1(e1,0)++an(en,0)+b1J(e1,0)++bnJ(en,0)
= (a1+ib1)(e1,0)++(an+ibn)(en,0),

where in the last line we used the definition of scalar multiplication in V. One checks that the set {(e1,0),,(en,0)} is linearly independent over , and therefore it is a basis for V. Hence

dimV=dimV.

3 Complexification is a functor

If V,W are real vector spaces and T:VW is an -linear map, we define

T:VW

by

T(v1,v2)=(Tv1,Tv2);

this is a -linear map. Setting ιV(v1,v2)=(v1,0) and ιW(w1,w2)=(w1,0), T:VW is the unique -linear map such that TιV=ιWT.22 2 See Keith Conrad’s, https://kconrad.math.uconn.edu/blurbs/linmultialg/complexification.pdf

Complexification is a functor from the category of real vector spaces to the category of complex vector spaces:

(idV)(v1,v2)=(idVv1,idVv2)=(v1,v2)=idV(v1,v2),

so (idV)=idV, and if S:UV and T:VS are -linear maps, then

(TS)(v1,v2)=(T(Sv1),T(Sv2))=T(Sv1,Sv2)=T(S(v1,v2)),

so (TS)=TS.

4 Complexifying a complex structure

If V is a real vector space and J:VV is a complex structure, then

(J)2(v1,v2) = J(Jv1,Jv2)
= (J2v1,J2v2)
= (-v1,-v2)
= -(v1,v2),

so (J)2=-idV. Let

Ei={wV:Jw=iw},E-i={wV:Jw=-iw}.

If wV, then one checks that

w-iJwEi,w+iJwE-i,

and

w=12(w-iJw)+12(w+iJw).

It follows that

V=EiE-i.

5 Complex structures, inner products, and symplectic forms

If V is a real vector space of odd dimension, then one can show that there is no linear map J:VV satisfying J2=-idV, i.e. there does not exist a complex structure for it. On the other hand, if V has even dimension, let

{e1,,en,f1,,fn}

be a basis for the real vector space V, and define J:VV by

Jej=fj,Jfj=-ej.

Then J:VV is a complex structure.

If V is a real vector space of dimension 2n with a complex structure J, let e10. Check that Je1span{e1}. If n>1, let

e2span{e1,Je1}.

Check that the set {e1,e2,Je1,Je2} is linearly independent. If n>2, let

e3span{e1,e2,Je1,Je2}.

Check that the set {e1,e2,e3,Je1,Je2,Je3} is linearly independent. If 2i<2n then there is some

ei+1span{e1,,ei,Je1,,Jei}.

I assert that

{e1,,en,Je1,,Jen}

is a basis for V.

Using the above basis {e1,,en,Je1,,Jen} for V, let fi=Jei, and define ,:V×V by

ei,ej=δi,j,fi,fj=δi,j,ei,fj=0,fi,ej=0.

Check that this is an inner product on the real vector space V. Moreover,

Jei,Jej=fi,fj=δi,j=ei,ej,

and

Jfi,Jfj=J2ei,J2ej=-ei,-ej=ei,ej=δi,j=fi,fj,

and

Jei,Jfj=fi,-ej=-fi,ej=0=ei,fj,

and

Jfi,Jej=-ei,fj=-ei,fj=0=fi,ej.

Hence for any v,wV,

Jv,Jw=v,w.

We say that the complex structure J is compatible with the inner product ,, i.e. J:(V,,)(V,,) is an orthogonal transformation.

A symplectic form on a real vector space V is a bilinear form ω:V×V such that ω(v,w)=-ω(w,v), and such that if ω(v,w)=0 for all w then v=0; we say respectively that ω is skew-symmetric and non-degenerate. If a real vector space V has a complex structure J, and , is an inner product on V that is compatible with J, define ω by

ω(v,w)=v,J-1w=v,-Jw=-v,Jw,

which is equivalent to

ω(v,Jw)=v,w.

Using that the inner product is compatible with J and that it is symmetric,

ω(v,w)=-v,Jw=-Jv,J2w=-Jv,-w=w,Jv=-ω(w,v),

so ω is skew-symmetric. If wV and ω(v,w)=0 for all vV, then

-v,Jw=0

for all vV, and thus Jw=0. Since J is invertible, w=0. Thus ω is nondegenerate. Therefore ω is a symplectic form on V.33 3 Using the basis {e1,,en,f1,,fn} for V, fi=Jei, we have ω(ei,fj)=-ei,Jfj=-ei,J2ej=-ei,-ej=ei,ej=δi,j, and ω(ei,ej)=-ei,Jej=-ei,fj=0,ω(fi,fj)=0. A basis {e1,,en,f1,,fn} for a symplectic vector space that satisfies these three conditions is called a Darboux basis. We have

ω(Jv,Jw)=-Jv,J2w=-J2v,J3w=--v,-Jw=-v,Jw=ω(v,w).

We say that J is compatible with the sympletic form ω, namely, J:(V,ω)(V,ω) is a symplectic transformation.

On the other hand, if V is a real vector space with symplectic form ω and J is a compatible complex structure, then ,:V×V defined by

v,w=ω(v,Jw)

is an inner product on V that is compatible with the complex structure J.

Suppose V is a real vector space with complex structure J:VV and that h:VJ×VJ is an inner product on the complex vector space VJ. Define g:V×V by44 4 The letter h refers to a Hermitian form, i.e. an inner product on a complex vector space, and the letter g refers to the usual notation for a metric on a Riemannian manifold.

g(v1,v2)=12(h(v1,v2)+h(v1,v2)¯)=12(h(v1,v2)+h(v2,v1)).

It is straightforward to check that g is an inner product on the real vector space V. Similarly, define ω:V×V by

ω(v1,v2)=i2(h(v1,v2)-h(v1,v2)¯)=i2(h(v1,v2)-h(v2,v1)).

It is apparent that ω is skew-symmetric. If ω(v1,v2)=0 for all v1, then in particular ω(iv1,v1)=0, and so

h(iv1,v1)-h(v1,iv1)=0.

As h is a complex inner product,

ih(v1,v1)-i¯h(v1,v1)=0,

i.e.

2ih(v1,v1)=0,

and thus h(v1,v1)=0, which implies that v1=0. Therefore ω is nondegenerate, and thus ω is a symplectic form on the real vector space V. With these definitions of g and ω, for v1,v2VJ we have

h(v1,v2)=g(v1,v2)-iω(v1,v2),

which writes the inner product on the complex vector space VJ using an inner product on the real vector space V and a symplectic form on the real vector space V; note that VJ has the same set of elements as V. Moreover, for v1,v2V we have

ω(v1,Jv2) = i2(h(v1,Jv2)-h(Jv2,v1))
= i2(h(v1,iv2)-h(iv2,v1))
= i2(-ih(v1,v2)-ih(v2,v1))
= g(v1,v2).

5.1 Tensor products

Here we give another presentation of the complexification of a real vector space, this time using tensor products of real vector spaces. If you were satisfied by the first definition you don’t need to read this one; read this either if you are curious about another way to define complexification, if you want to see a pleasant application of tensor products, or if you didn’t like the first definition. Let V be a real vector space of dimension n. is a real vector space of dimension 2, and

V

is a real vector space of dimension 2n. If V has basis {e1,,en}, then V has basis {e11,,en1,e1i,,eni}. Since every element of V can be written uniquely in the form

v11+v2i,v1,v2V,

one often writes

VViV;

here iV is a real vector space that is isomorphic to V.

The complexification of V is the complex vector space V whose set of elements is V, with the same addition as the real vector space V, and with scalar multiplication defined by

α(vβ)=v(αβ),vV,α,β.

Let vV. Using the basis of the real vector space V, there exist some

a1,,an,b1,,bn

such that

v = a1e11++anen1+b1e1i++bneni
= e1(a1+ib1)++en(an+ibn)
= (a1+ib1)e11++(an+ibn)en1,

where in the last line we used the definition of scalar multiplication in V. One checks that the {e11,,en1} is linearly independent over , and hence that it is a basis for the complex vector space V, so V has dimension n over .

If V and W are real vector spaces and T:VW is a linear map, define T:VW by

T(vz)=(Tv)z.

With this definition of T, one can check that complexification is a functor from the category of real vector spaces to the category of complex vector spaces.

6 Decomplexification

If V is a complex vector space, let V be the real vector space whose set of elements is V, in which addition is the same as addition in V, and in which scalar multiplication is defined by

av=(a+0i)v,a.

Because V is a complex vector space, it is apparent that V is a real vector space with this scalar multiplication. We call V the decomplexification of the complex vector space V.

If V has basis {e1,,en} and vV, then there are a1+ib1,,an+ibn such that

v=(a1+ib1)e1++(an+ibn)en=a1e1++anen+b1(ie1)++bn(ien).

One checks that

e1,,en,ie1,,ien

are linearly independent over , and hence are a basis for the real vector space V. Thus,

dimV=2dimV.

If V is a complex vector space and T:VV is a -linear map, define T:VV by

Tv=Tv.

Because T is -linear it follows that T is -linear. Decomplexification is a functor from the category of complex vector spaces to the category of real vector spaces. Since decomplexification is defined simply by ignoring the fact that V is closed under multiplication by complex scalars and only using real scalars, decomplexification is called a forgetful functor

7 Complex conjugation in complexified vector spaces

If V is a real vector space, define σ:VV by

σ(v1,v2)=(v1,-v2).

We call σ complex conjugation in V. We have σσ=idV. If T:VV is a -linear map, define Tσ:VV by

Tσ(w)=σ(Tσ(w)).

Tσ is a -linear map. It is a fact that if T:VV is -linear, then Tσ=T if and only if there is some -linear S:VV such that T=S. In words, a linear map on the complexification of a real vector space is equal to its own conjugate if and only if it is the complexification of a linear map on the real vector space.

The following are true statements:55 5 These are exercises from V. I. Arnold’s Ordinary differential equations, p. 122, §18.4, in Richard A. Silverman’s translation. (n=(n))

  • If T:nn is a linear map, then

    exp(T)=exp(T),

    and

    exp(T)σ=exp(Tσ).
  • If T:nn is a linear map, then

    exp(T)=exp(T).
  • If T:nn is a linear map, then

    detT=|detT|2,

    and

    detTσ=detT¯.
  • If T:nn is a linear map, then

    detT=detT.
  • If T:nn is a linear map, then

    Tr(T)=TrT+TrTσ,

    and

    TrTσ=TrT¯.
  • If T:nn is a linear map, then

    TrT=TrT.

8 Linear ordinary differential equations over

Let A be an n×n matrix over . The solution of the initial value problem

z(t)=Az(t),z(0)=z0n,

is z(t)=exp(At).

If A has n distinct eigenvalues λ1,,λn, then, with

Eλ={zn:Az=λz},

we have

n=Eλ1++Eλn,

where each Eλk has dimension 1. For zEλk,

exp(At)z=m=0tmAmzm!=m=0tmλkmzm!=eλktz.

Let ξkEλk be nonzero, 1kn. They are a basis for n, so there are ck such that

z0=k=1nckξk.

Then

z(t)=exp(At)z0=exp(At)k=1nckξk=k=1nckexp(At)ξk=k=1nckeλktξk.

Suppose that A is an n×n matrix over , that z0n, that Aσ=A and that σ(z0)=z0. The solution of the initial value problem

z(t)=Az(t),z(0)=z0,

is z(t)=exp(At)z0. We have, as exp(At)σ=exp((At)σ)=exp(At),

σ(z(t))=σ(exp(At)z0)=σ(exp(At)σ(z0))=exp(At)σz0=exp(At)z0=z(t).

Therefore, if Aσ=A and σ(z0)=z0, then σ(z(t))=z(t) for all t.

9 Linear ordinary differential equations over

Let A be an n×n matrix over and let x0n. Let z0=(x0,0)n=(n), and let z(t)=(z1(t),z2(t)) be the solution of the initial value problem

z(t)=Az(t),z(0)=z0n.

As A is the complexification of a real linear map, (A)σ=A, and

σ(z0)=σ(x0,0)=(x0,-0)=(x0,0)=z0,

so σ(z(t))=z(t), i.e. (z1(t),z2(t))=(z1(t),-z2(t)), so z2(t)=0 for all t. But z(t)=(z1(t),z2(t)) and Az(t)=(Az1(t),Az2(t)), so

z1(t)=Az1(t)

for all t. Also, z(0)=(z1(0),z2(0)) and z(0)=z0=(x0,0), so z1(0)=x0. Therefore, x(t)=z1(t) is the solution of the initial value problem

x(t)=Ax(t),x(0)=x0.

Thus, to solve an initial value problem in n we can complexify it, solve the initial value problem in n, and take the first entry of the solution of the complex initial value problem.

If A is an n×n matrix over , let

det(A-xI)=k=0nakxk,ak,

its characteristic polynomial. The Cayley-Hamilton theorem states that

k=0nakAk=0.

Taking the complexification of this gives

k=0nak(A)k=0.

It follows that the roots of det(A-xI) are the same as the roots of det(A-xI). A complex root of det(A-xI) is not an eigenvalue of A:22, but is indeed an eigenvalue of A:22, so the roots of the characteristic polynomial of A are the eigenvalues of A.

10 Linear ordinary differential equations in 2

Let A be a 2×2 matrix over .66 6 This section follows Arnold, p. 132, §20.3. Suppose that the roots of the characteristic polynomial

det(A-xI)=detA-xTrA+x2

are λ,λ¯, i.e. that the roots of the characteristic polynomial are complex conjugate. Let λ=α+iω, ω0.77 7 Define J:22 by J=1ω(A-αI). We have J2=1ω2(A2-2αA+α2I). By the Cayley-Hamilton theorem, IdetA-ATrA+A2=0, so Iλλ¯-A(λ+λ¯)+A2=0, and written using λ=α+iω this is I(α2+ω2)-2αA+A2=0. Hence J2=-I, so J=1ω(A-αI) is a complex structure on 2. λ is an eigenvalue for A, so let A(v1,v2)=λ(v1,v2), (v1,v2)0. Furthermore,

σ(A(v1,v2))=σ(λ(v1,v2)),

so

(A)σσ(v1,v2)=λ¯σ(v1,v2),

hence, as (A)σ=A,

A(v1,-v2)=λ¯(v1,-v2).

Therefore (v1,-v2) is an eigenvector of A with eigenvalue λ¯λ, so (v1,-v2) and (v1,v2) are linearly independent over . If a1v1+a2v2=0, a1,a2, then

(a12-ia22)(v1,v2)+(a12+ia22)(v1,-v2)=0,

from which it follows that a1,a2=0. Therefore v1,v22 are linearly independent over .

We have

(α+iω)(v1,v2)=(αv1-ωv2,αv1+ωv2),

and

A(v1,v2)=(Av1,Av2),

so

Av1=αv1-ωv2,Av2=αv1+ωv2,

and hence

A(v1v2)=(αv1-ωv2αv1+ωv2)=(v1v2)(αω-ωα).

Therefore

A=(v1v2)(αω-ωα)(v1v2)-1.