Difference between revisions of "Aufgaben:Problem 4"

From Ferienserie MMP2
Jump to: navigation, search
(Added Solution)
m ((a): Put a \(\forall\) in front of the according statement)
 
(31 intermediate revisions by 8 users not shown)
Line 1: Line 1:
 
==Note==
 
==Note==
  
Check out the document Lie-Gruppen, Bsp. 3.2 for part a) here: [https://www.dropbox.com/sh/su8ja1eynb449nr/AABIiXtiB6SNdwjjWYtYN8Tua?dl=0]. Just be careful, he writes \(\mathfrak{sl} (2, \mathbf{R}) = \{ A \in \mathbf{R}^{d \times d} : \mathrm{Tr}(A) = 1 \} \) while the trace should be zero. Just a typo.  
+
Check out the document Lie-Gruppen, Bsp. 3.2 for part a) here: [https://www.dropbox.com/sh/su8ja1eynb449nr/AABIiXtiB6SNdwjjWYtYN8Tua?dl=0]. Just be careful, he writes \(\mathfrak{sl} (2, \mathbf{R}) = \{ A \in \mathbf{R}^{d \times d} : \mathrm{Tr}(A) = 1 \} \) while the trace should be zero. Just a typo.
 +
 
 +
 
 +
Here is the solution by Grégoire: [[Media:Ex._4.pdf]]
  
 
==Taks==
 
==Taks==
Line 8: Line 11:
  
 
(b) show that
 
(b) show that
$$H = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right),\ E_+= \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right),\ E_- \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)$$
+
$$H = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right),\ E_+= \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right),\ E_- \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)$$
  
 
form a basis of \(\mathfrak{g} = \mathfrak{sl} (2, \mathbf{R}) \) as a vector space over \(\mathbb{R}\) and show that they satisfy the relations:
 
form a basis of \(\mathfrak{g} = \mathfrak{sl} (2, \mathbf{R}) \) as a vector space over \(\mathbb{R}\) and show that they satisfy the relations:
Line 21: Line 24:
  
 
===(a)===
 
===(a)===
$$SL(2,\mathbb{R}) = \{A \in Mat(2,\mathbb{R}) |\ det(A) = 1\}$$
+
$$SL(2,\mathbb{R}) = \{A \in \mathrm{Mat}(2,\mathbb{R}) |\ \mathrm{det}(A) = 1\}$$
$$\mathfrak{sl}(2,\mathbb{R}) = \{\dot{\gamma}\ (0)\ | \gamma(-\epsilon, \epsilon)\rightarrow SL(2,\mathbb{R}),\ \gamma(0) = \mathbb{I}\}$$
+
$$\mathfrak{sl}(2,\mathbb{R}) = \{\dot{\gamma}\ (0)\ | \gamma : (-\epsilon, \epsilon)\rightarrow SL(2,\mathbb{R}),\ \gamma(0) = \mathbb{I}\}$$
  
let \(\gamma\) be a curve in \(SL(2\mathbb{R})\) such that \(\gamma(0) = \mathbb{I}\)
+
let \(\gamma\) be a curve in \(SL(2,\mathbb{R})\) such that \(\gamma(0) = \mathbb{I}\)
  
 
$$ \gamma(t) = \left( \begin{matrix} a & b\\ c & d\end{matrix}\right)$$
 
$$ \gamma(t) = \left( \begin{matrix} a & b\\ c & d\end{matrix}\right)$$
  
$$ \Rightarrow det(\gamma(t)) = (ad -bc) = 1$$  
+
$$ \Rightarrow \forall t \in [-\epsilon, \epsilon]: \: \mathrm{det}(\gamma(t)) = (ad -bc) = 1.$$  
  
\( \forall t \in \mathbb{R} \). We take the derivative to find the conditions for the elements of \(\mathfrak{sl}(2,\mathbb{R}) \):
+
We take the derivative to find the conditions for the elements of \(\mathfrak{sl}(2,\mathbb{R}) \):
  
$$ \frac{d}{dt} det(\gamma(t)) = 0$$
+
$$ \frac{d}{dt} \mathrm{det}(\gamma(t)) = 0$$
  
 
$$\dot{a}d + a\dot{d} - \dot{b}c - b\dot{c} =  0$$
 
$$\dot{a}d + a\dot{d} - \dot{b}c - b\dot{c} =  0$$
Line 38: Line 41:
 
Consider the following:
 
Consider the following:
  
$$tr(\gamma(t)^{-1}\dot{\gamma}(t))$$
+
(Note that \(\gamma(t)\) is always invertible because \(\mathrm{det}(\gamma(t))= 1 \neq 0\))
  
$$ = tr \left( \frac{1}{ad-bd} \left( \begin{matrix} d & -b\\-c & a\end{matrix}\right) \left( \begin{matrix} \dot{a} & \dot{b}\\ \dot{c} & \dot{d}\end{matrix}\right)\right)$$
+
$$\mathrm{tr}(\gamma(t)^{-1}\dot{\gamma}(t))$$
  
$$  = \frac{1}{ad-bd} tr\left( \begin{matrix} d \dot{a} + b \dot{c} & \dots\\ \dots & -c \dot{b} -a \dot{d}\end{matrix}\right) = \frac{1}{ad-bd}(d\dot{a}+ b\dot{c} -c\dot{b}- a\dot{d})$$
+
$$ = \mathrm{tr} \left( \frac{1}{ad-bc} \left( \begin{matrix} d & -b\\-c & a\end{matrix}\right) \left( \begin{matrix} \dot{a} & \dot{b}\\ \dot{c} & \dot{d}\end{matrix}\right)\right)$$
 +
 
 +
$$  = \frac{1}{ad-bc} \mathrm{tr}\left( \begin{matrix} d \dot{a} - b \dot{c} & \dots\\ \dots & -c \dot{b} +a \dot{d}\end{matrix}\right) = \frac{1}{ad-bc}(d\dot{a}- b\dot{c} -c\dot{b}+ a\dot{d})$$
  
 
we conclude:
 
we conclude:
  
$$ det(\gamma(t)) =0 \Leftrightarrow tr(\gamma(t)^{-1}\dot{\gamma}(t))=0 $$
+
$$ \frac{d}{dt} \mathrm{det}(\gamma(t)) =0 \Leftrightarrow \mathrm{tr}(\gamma(t)^{-1}\dot{\gamma}(t))=0 $$
  
 
we are interested in the point \(t = 0\):
 
we are interested in the point \(t = 0\):
  
$$tr(\gamma(0)^{-1}\dot{\gamma}(0)) = tr(\mathbb{I}^{-1}\dot{\gamma}(0)) = tr(\dot{\gamma}(0)) \overset{!}{=} 0$$
+
$$\mathrm{tr}(\gamma(0)^{-1}\dot{\gamma}(0)) = \mathrm{tr}(\mathbb{I}^{-1}\dot{\gamma}(0)) = \mathrm{tr}(\dot{\gamma}(0)) \overset{!}{=} 0$$
 +
 
 +
$$\Rightarrow \mathfrak{sl}(2,\mathbb{R}) \subset \{A \in \mathrm{Mat}(2,\mathbb{R}) | \mathrm{tr}(A) = 0\}$$
 +
 
 +
Now for \(\supset\): let \(A\in \mathrm{Mat}(2,\mathbb{R})\) such that \(\mathrm{tr}(A) = 0\). We define a curve \(\gamma: (-\epsilon,\epsilon) \rightarrow SL(2,\mathbb{R})\) the following way: \( \gamma(t) =e^{tA}\). We check that  \( \dot{\gamma}(0) = A\) and \(\gamma(0) = e^{0A} = \mathbb{I}\Rightarrow \dot{\gamma}(0) \in  \mathfrak{sl}(2,\mathbb{R})\). We only have to show that the curve actually is a map onto \(SL(2,\mathbb{R})\), to do this we calculate the determinante:
 +
 
 +
$$ \mathrm{det}(\gamma(t)) = \mathrm{det}(e^{tA}) = e^{\mathrm{tr}(tA)} = e^{t \cdot \mathrm{tr}(A)} =e^0 = 1$$
  
$$\Rightarrow \mathfrak{sl}(2,\mathbb{R}) = \{A \in Mat(2,\mathbb{R}) | tr(A) = 0\}$$
+
$$\Rightarrow \mathfrak{sl}(2,\mathbb{R}) = \{A \in \mathrm{Mat}(2,\mathbb{R}) | \mathrm{tr}(A) = 0\}$$
  
 
===(b)===
 
===(b)===
Line 62: Line 73:
 
$$ = aH + bE_+ + cE_-$$
 
$$ = aH + bE_+ + cE_-$$
  
left to show: \(aH + bE_+ + cE_-\) are linear independent. let \(a,b,c,d\in \mathbb{R}\) and:
+
left to show: \(aH + bE_+ + cE_-\) are linear independent. let \(a,b,c\in \mathbb{R}\) and:
  
 
$$ aH + bE_+ + cE_- = 0$$
 
$$ aH + bE_+ + cE_- = 0$$
Line 68: Line 79:
 
$$\left(\begin{matrix} a & 0\\ 0 & -a \end{matrix}\right) + \left(\begin{matrix} 0 & b\\ 0 & 0 \end{matrix}\right) + \left(\begin{matrix} 0 & 0\\ c & 0 \end{matrix}\right) = 0$$
 
$$\left(\begin{matrix} a & 0\\ 0 & -a \end{matrix}\right) + \left(\begin{matrix} 0 & b\\ 0 & 0 \end{matrix}\right) + \left(\begin{matrix} 0 & 0\\ c & 0 \end{matrix}\right) = 0$$
  
$$\left(\begin{matrix} a & b\\ c & -a \end{matrix}\right) = \left(\begin{matrix} 0 & 0\\ 0 & 0 \end{matrix}\right) \Rightarrow a=b=c=d = 0$$
+
$$\left(\begin{matrix} a & b\\ c & -a \end{matrix}\right) = \left(\begin{matrix} 0 & 0\\ 0 & 0 \end{matrix}\right) \Rightarrow a=b=c = 0$$
  
 
Now we calculate the commutators:
 
Now we calculate the commutators:
Line 74: Line 85:
 
$$[H,E_+] = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right)\left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)\left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right) = \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)  - \left(\begin{matrix} 0 & -1\\ 0 & 0 \end{matrix}\right) = 2E_+$$
 
$$[H,E_+] = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right)\left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)\left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right) = \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)  - \left(\begin{matrix} 0 & -1\\ 0 & 0 \end{matrix}\right) = 2E_+$$
  
$$[H,E_-] = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right)\left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)\left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right) = \left(\begin{matrix} 0 & 0\\ -1 & 0 \end{matrix}\right)  - \left(\begin{matrix} 0 & 0\\ -1 & 0 \end{matrix}\right) = -2E_-$$
+
$$[H,E_-] = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right)\left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)\left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right) = \left(\begin{matrix} 0 & 0\\ -1 & 0 \end{matrix}\right)  - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) = -2E_-$$
  
 
$$[E_+,E_-] = \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)\left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)\left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) = \left(\begin{matrix} 1 & 0\\ 0 & 0 \end{matrix}\right)  - \left(\begin{matrix} 0 & 0\\ 0 & 1 \end{matrix}\right) = H$$
 
$$[E_+,E_-] = \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)\left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)\left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) = \left(\begin{matrix} 1 & 0\\ 0 & 0 \end{matrix}\right)  - \left(\begin{matrix} 0 & 0\\ 0 & 1 \end{matrix}\right) = H$$
  
We conclude that \([\mathfrak{g}, \mathfrak{g}] \) the span of all commutators with elements in \(\mathfrak{g}\) is again equal to \(\mathfrak{g}\): With the linearity and antisymmetry of the commutator (And that fact that \([A,A] = 0\)) we see that every element in \([\mathfrak{g}, \mathfrak{g}] \) decomposes into a linear combination of the above commutators and thus into a linear combination of \(H, E_+, E_-\). As \(H, E_+, E_-\) clearly are also elements of \([\mathfrak{g}, \mathfrak{g}] \), they form a basis of \([\mathfrak{g}, \mathfrak{g}] \) (linear independence already established). Because \([\mathfrak{g}, \mathfrak{g}] \) and \(\mathfrak{g}\) share a basis they are equal.
+
We conclude that \([\mathfrak{g}, \mathfrak{g}] \) is again equal to \(\mathfrak{g}\): With the linearity and antisymmetry of the commutator (And that fact that \([A,A] = 0\)) we see that every element in \([\mathfrak{g}, \mathfrak{g}] \) decomposes into a linear combination of the above commutators and thus into a linear combination of \(H, E_+, E_-\). Thus  \(H, E_+, E_-\) form a basis of \([\mathfrak{g}, \mathfrak{g}] \) (linear independence already established). Because \([\mathfrak{g}, \mathfrak{g}] \) and \(\mathfrak{g}\) share a basis they are equal.

Latest revision as of 13:52, 26 July 2015

Note

Check out the document Lie-Gruppen, Bsp. 3.2 for part a) here: [1]. Just be careful, he writes \(\mathfrak{sl} (2, \mathbf{R}) = \{ A \in \mathbf{R}^{d \times d} : \mathrm{Tr}(A) = 1 \} \) while the trace should be zero. Just a typo.


Here is the solution by Grégoire: Media:Ex._4.pdf

Taks

(a) Compute the Lie algebra \(\mathfrak{sl}(2,\mathbb{R})\) of \(SL(2,\mathbb{R})\)

(b) show that $$H = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right),\ E_+= \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right),\ E_- \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)$$

form a basis of \(\mathfrak{g} = \mathfrak{sl} (2, \mathbf{R}) \) as a vector space over \(\mathbb{R}\) and show that they satisfy the relations:

$$[H,E_+] = 2E_+$$ $$[H,E_-] = -2E_-$$ $$[E_+,E_-] = H$$

Here \([,]\) denotes the matrix commutator. Conclude that \([\mathfrak{g}, \mathfrak{g}] = \mathfrak{g}\). Here, the set \([\mathfrak{g}, \mathfrak{g}] \) is de defined as the span of all commutators between elements of \(\mathfrak{g}\)

Solution

(a)

$$SL(2,\mathbb{R}) = \{A \in \mathrm{Mat}(2,\mathbb{R}) |\ \mathrm{det}(A) = 1\}$$ $$\mathfrak{sl}(2,\mathbb{R}) = \{\dot{\gamma}\ (0)\ | \gamma : (-\epsilon, \epsilon)\rightarrow SL(2,\mathbb{R}),\ \gamma(0) = \mathbb{I}\}$$

let \(\gamma\) be a curve in \(SL(2,\mathbb{R})\) such that \(\gamma(0) = \mathbb{I}\)

$$ \gamma(t) = \left( \begin{matrix} a & b\\ c & d\end{matrix}\right)$$

$$ \Rightarrow \forall t \in [-\epsilon, \epsilon]: \: \mathrm{det}(\gamma(t)) = (ad -bc) = 1.$$

We take the derivative to find the conditions for the elements of \(\mathfrak{sl}(2,\mathbb{R}) \):

$$ \frac{d}{dt} \mathrm{det}(\gamma(t)) = 0$$

$$\dot{a}d + a\dot{d} - \dot{b}c - b\dot{c} = 0$$

Consider the following:

(Note that \(\gamma(t)\) is always invertible because \(\mathrm{det}(\gamma(t))= 1 \neq 0\))

$$\mathrm{tr}(\gamma(t)^{-1}\dot{\gamma}(t))$$

$$ = \mathrm{tr} \left( \frac{1}{ad-bc} \left( \begin{matrix} d & -b\\-c & a\end{matrix}\right) \left( \begin{matrix} \dot{a} & \dot{b}\\ \dot{c} & \dot{d}\end{matrix}\right)\right)$$

$$ = \frac{1}{ad-bc} \mathrm{tr}\left( \begin{matrix} d \dot{a} - b \dot{c} & \dots\\ \dots & -c \dot{b} +a \dot{d}\end{matrix}\right) = \frac{1}{ad-bc}(d\dot{a}- b\dot{c} -c\dot{b}+ a\dot{d})$$

we conclude:

$$ \frac{d}{dt} \mathrm{det}(\gamma(t)) =0 \Leftrightarrow \mathrm{tr}(\gamma(t)^{-1}\dot{\gamma}(t))=0 $$

we are interested in the point \(t = 0\):

$$\mathrm{tr}(\gamma(0)^{-1}\dot{\gamma}(0)) = \mathrm{tr}(\mathbb{I}^{-1}\dot{\gamma}(0)) = \mathrm{tr}(\dot{\gamma}(0)) \overset{!}{=} 0$$

$$\Rightarrow \mathfrak{sl}(2,\mathbb{R}) \subset \{A \in \mathrm{Mat}(2,\mathbb{R}) | \mathrm{tr}(A) = 0\}$$

Now for \(\supset\): let \(A\in \mathrm{Mat}(2,\mathbb{R})\) such that \(\mathrm{tr}(A) = 0\). We define a curve \(\gamma: (-\epsilon,\epsilon) \rightarrow SL(2,\mathbb{R})\) the following way: \( \gamma(t) =e^{tA}\). We check that \( \dot{\gamma}(0) = A\) and \(\gamma(0) = e^{0A} = \mathbb{I}\Rightarrow \dot{\gamma}(0) \in \mathfrak{sl}(2,\mathbb{R})\). We only have to show that the curve actually is a map onto \(SL(2,\mathbb{R})\), to do this we calculate the determinante:

$$ \mathrm{det}(\gamma(t)) = \mathrm{det}(e^{tA}) = e^{\mathrm{tr}(tA)} = e^{t \cdot \mathrm{tr}(A)} =e^0 = 1$$

$$\Rightarrow \mathfrak{sl}(2,\mathbb{R}) = \{A \in \mathrm{Mat}(2,\mathbb{R}) | \mathrm{tr}(A) = 0\}$$

(b)

let \(A \in \mathfrak{g}\)

$$\Rightarrow A = \left(\begin{matrix} a & b\\ c & -a \end{matrix}\right) = \left(\begin{matrix} a & 0\\ 0 & -a \end{matrix}\right) + \left(\begin{matrix} 0 & b\\ 0 & 0 \end{matrix}\right) + \left(\begin{matrix} 0 & 0\\ c & 0 \end{matrix}\right)$$

$$ = aH + bE_+ + cE_-$$

left to show: \(aH + bE_+ + cE_-\) are linear independent. let \(a,b,c\in \mathbb{R}\) and:

$$ aH + bE_+ + cE_- = 0$$

$$\left(\begin{matrix} a & 0\\ 0 & -a \end{matrix}\right) + \left(\begin{matrix} 0 & b\\ 0 & 0 \end{matrix}\right) + \left(\begin{matrix} 0 & 0\\ c & 0 \end{matrix}\right) = 0$$

$$\left(\begin{matrix} a & b\\ c & -a \end{matrix}\right) = \left(\begin{matrix} 0 & 0\\ 0 & 0 \end{matrix}\right) \Rightarrow a=b=c = 0$$

Now we calculate the commutators:

$$[H,E_+] = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right)\left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)\left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right) = \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & -1\\ 0 & 0 \end{matrix}\right) = 2E_+$$

$$[H,E_-] = \left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right)\left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)\left(\begin{matrix} 1 & 0\\ 0 & -1 \end{matrix}\right) = \left(\begin{matrix} 0 & 0\\ -1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) = -2E_-$$

$$[E_+,E_-] = \left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right)\left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 1 & 0 \end{matrix}\right)\left(\begin{matrix} 0 & 1\\ 0 & 0 \end{matrix}\right) = \left(\begin{matrix} 1 & 0\\ 0 & 0 \end{matrix}\right) - \left(\begin{matrix} 0 & 0\\ 0 & 1 \end{matrix}\right) = H$$

We conclude that \([\mathfrak{g}, \mathfrak{g}] \) is again equal to \(\mathfrak{g}\): With the linearity and antisymmetry of the commutator (And that fact that \([A,A] = 0\)) we see that every element in \([\mathfrak{g}, \mathfrak{g}] \) decomposes into a linear combination of the above commutators and thus into a linear combination of \(H, E_+, E_-\). Thus \(H, E_+, E_-\) form a basis of \([\mathfrak{g}, \mathfrak{g}] \) (linear independence already established). Because \([\mathfrak{g}, \mathfrak{g}] \) and \(\mathfrak{g}\) share a basis they are equal.