数式を処理中: 77%

Introduction

I have read KASAHARA [KAS] with @kyotomathmath. This book brings me into consideration on the definition of differentiations z,¯z. Let’s take a look at what happens in discussing differentiations of complex functions.

The organization of this article follows that of [KAS]; hence the chapter 1 of this corresponds to the chapter 1 of [KAS], etc.

What are holomorphic functions?

Preliminaries

Let’s define (formal) partial differentials z,¯z for complex functions.

Fix a field k, and consider an algebra A over k.

A linear map D:AA is called a derivation is D(ab)=D(a)b+aD(b)(a,bA).

Denote the set of all derivations on A by DerkA: DerkA:={D:AA:derivation}.

When there is no ambiguity, we denote DerkA simply by DerA.

It’s easy to show that DerAgl(A) is a Lie subalgebra of gl(A).

Let k=R and A=i=1Ci(D), where DRn and Ci(D):={f:DR:Ci class}.

The partial differentials i:AAfxif are derivations on A.

The following lemma is useful to construct a new derivation from old ones.

Let k be a field, A,B algebras /k.

  1. AB is also an algebra /k.
  2. DDerA, EDerB, DEDer(AB).
  3. The map DerADerBDer(AB)(D,E)DE is an injective Lie algebra hom.

This lemma can easily be verified by straightforward calculations.

Again, consider the case k=R, DRn, A=Ci(D). Since there exists an isomorphism (Ci(D))mCi(D,Rm) as vector spaces /R, where Ci(D,Rm):={f:DRm:Ci class}, and since Am is an algebra /R by the above lemma, Ci(D,Rm) can be regarded as an algebra /R.

Then, e.g. i1imDer(Ci(D,Rm)) (1ijn).

Next, suppose we have an extension of fields kK. Any vector space V /k can be extended to the vector space VkK /K. Explicitly, the scalar multiplication on VkK is given by μ(vλ):=v(μλ)(vV,λ,μK).

Suppose, furthermore, A is an algebra /k and consider the vector space AkK /K. This is now an algebra /K; indeed, the ring hom kA induces a ring hom K(kkK)AkK, which makes AkK into an algebra /K.

Let kK be extension of fields, A algebra /k, AkK algebra /K.

  1. fEndkA, f1EndK(AkK).
  2. DDerkA, D1DerK(AkK).

i) It’s obvious that f1 preserves addition.

For any xA, λ,μK, (f1)(μ(xλ))=f(x)μλ=μ(f(x)λ)=μ(f1)(xλ).

ii) We’ve already seen that D1 is K-linear. To see that is is also a derivation, note that, for x,yA, λ,μK, (xλ)(yμ)=(xy)(λμ). Thus, for any x,yA, λ,μK, we have (D1)((xλ)(yμ))=D(xy)λμ=D(x)yλμ+xD(y)λμ=((D1)(xλ))(yμ)+(xλ)((D1)(yμ)).

In the following discussion, let’s focus on the special case RC, the degree of whose extension is 2.

First, let’s describe the complexification ARC more explicitly. Since C=RiR, we can write, for any vector space V /R, VRC=(VRR)(VRiR)=ViV as vector spaces /R.

If A is an algebra /R, let J:ARCARC:R-linear,azaiz, so that J2=id. Therefore we have a ring hom CEndR(ARC)x+iyxid+yJ. Endowed with this structure map, ARC becomes an algebra /C as described above.

For a linear map f:VW between vector spaces /R, the C-linear map f1:VRCWRC is now given by the formula ViVWiWv+iwf(v)+if(w); hence denoted by ff:VRCWRC.

Let DR2, A=Ci(D), A2=Ci(D,R2).

By the above lemma, jjDerA2 (1j2), regarded as an algebra /C. Since DerA2 is a vector space /C, z:=12(xxiyy),¯z:=12(xx+iyy) are also derivations on A2: z,¯zDerA2. These satisfies xx=z+¯z,yy=i(z¯z), which justifies the notation “z, ¯z”.

ARC is equipped with another inmortant endmorphism SEndR(ARC) represented by either S:ARCARCaza¯z, S:AiAAiAa+ibaib. In general, an endmorphism SEndRV of a vector space V /R is called an involution if S2=id but Sid.

The following lemma shows the compatibility of the involution S on ARC with that on C.

Let A,B be algebras /R, SA,SB involutions on A,B, resp., given above.

  1. f:AB: linear, SB(f1)=(f1)SA.
  2. F1,F2,G1,G2:ARCBRC: C-linear, λC, if SBFj=GjSA(j=1,2), then SB(F1+F2)=(G1+G2)SA,SB(λF1)=(¯λG1)SA

i) aA, zC, we have (SB(f1))(az)=SB(f(a)z)=f(a)¯z,((f1)SA)(az)=(f1)(a¯z)=f(a)¯z; hence SB(f1)=(f1)SA.

ii) THe first assertion is obvious as SB(F1+F2)=SBF1+SBF2=G1SA+G2SA=(G1+G2)SA.

The second follows from that aA, zC, we have SB(λF1)(az)=SB(F1(aλz))=G1(SA(aλz))=G1(a¯λ¯z)=¯λ(G1SA(az)).

Let DR2, A=Ci(D), A2=Ci(D,R2).

Since x,yDerAEndRA, by the above lemma, SA¯z=zSA, which is expressed effectively as ¯¯zf=z¯f(fA), where ¯f denotes the map ¯f(x,y):=¯f(x,y).

The lemma 1.1.1 of [KAS] is now obvious from above consideration.


Now that we define z,¯z by z:=12(xxiyy),¯z:=12(xx+iyy), and from this follow the relations zz=1,z¯z=0,¯zz=0,¯z¯z=1.

Conversely, if we start with zz=1, etc., then we can derive the formula for z,¯z, or equivalently, \begin{align*} \xder \oplus \xder &= \zder + \bder, \\ \yder \oplus \yder &= \iunit (\zder - \bder). \end{align*}

Let \dcal \subset \R^2, A = \sum C^i (\dcal), A^2 = \sum C^i (\dcal, \R^2), as above, and \zder, \bder \in \Der A^2 satisfy \begin{align*} \zder z = 1, &\quad \zder \zbar = 0, \\ \bder z = 0, &\quad \bder \zbar = 1. \end{align*}

For x = (z + \zbar) / 2, y = (z - \zbar) / 2 \iunit, we have \begin{align*} \zder x = \frac 12, &\quad \zder y = \frac 1{2\iunit}, \\ \bder x = \frac 12, &\quad \bder y = - \frac 1{2\iunit}. \end{align*}

By the composition laws, which we haven’t proved yet, we want \begin{align*} \xder z = 1, &\quad \xder \zbar = 1, \\ \yder z = \iunit, &\quad \yder \zbar = -\iunit, \end{align*} and \forall f \in A^2, \begin{align*} \xder f &= (\xder z) (\zder f) + (\xder \zbar) (\bder f) \\ &= \zder f + \bder f, \\ \yder f &= (\yder z) (\zder f) + (\yder \zbar) (\bder f) \\ &= \iunit \zder f - \iunit \bder f. \end{align*}

So, it’s reasonable to define \begin{align*} \xder &:= \zder + \bder, \\ \yder &:= \iunit (\zder - \bder). \end{align*}

Then, are these operators agree with \xder \oplus \xder and \yder \oplus \yder, resp.?

For definiteness, we denote the new operators instead by \begin{align*} \xder' &:= \zder + \bder, \\ \yder' &:= \iunit (\zder - \bder). \end{align*} What we must show is then \forall f \in A, \begin{align*} \xder' f = \xder f, \\ \yder' f = \yder f. \end{align*}

First note that \begin{align*} \xder' x = 1, &\quad \xder' y = 0, \\ \yder' x = 0, &\quad \yder' y = 1, \end{align*} We can see the linear independence of \xder', \yder' from these relations: for any \lambda, \mu \in \C, if \lambda \xder' + \mu \yder' = 0, then applying x to the both sies gives \lambda = (\lambda \xder' + \mu \yder') (x) = 0, and similarly we obtain \mu = 0.

Let B := \{ f \in A \colon \text{polynomial} \} = \R [x, y].

Since \xder', \yder' \in \Der A^2, we have \xder' = \xder and \yder' = \yder on B (using the defining properties of derivations). So, if \dcal is bounded, then, by the Stone-Weierstrass’ theorem, B is dense in A; hence \xder' = \xder and \yder' = \yder on A.

Otherwise, since \R^2 is locally compact, there exists a family (\bcal_j)_{j \in I} of bounded sets indexed by some set I such that \dcal = \bigcup_{j \in I} \bcal_j. Then, \xder' = \xder and \yder' = \yder on each \sum C^i (\bcal_j), and hence on A because \sum C^i (\dcal) \subset \bigcup_j \sum C^i (\bcal_j).