From Simple Groups to Quantum Field Theory

Click For Summary

Discussion Overview

The discussion explores the connections between simple groups, specifically SU(2), and quantum field theory, focusing on mathematical formulations involving exponential functions, derivatives, and algebraic structures. Participants delve into various mathematical concepts such as the Kronecker delta, permutation symbols, and their applications in physics.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • Garrett discusses the use of the exponential function to represent rotations in three dimensions using bivectors, proposing a specific formulation involving the angle and unit bivectors.
  • Another participant relates the generators of the Lie algebra su(2) to Euler's formula, suggesting that the group elements can be expressed through exponentiation of these generators.
  • A participant introduces the Levi-Civita symbol, detailing its properties and relationships to the Kronecker delta, emphasizing its role in vector calculus.
  • Discussion on the Kronecker delta includes its interpretation as a discrete delta function and its identities in three-dimensional space.
  • Participants explore various types of derivatives, including partial and total derivatives, and their significance in mathematical analysis and physics.
  • Chain rule and derivatives of exponential and logarithmic functions are presented, with participants sharing formulas and interpretations.
  • There is a mention of the derivative's role as a linear approximation of functions, although this point remains incomplete.

Areas of Agreement / Disagreement

Participants express a range of viewpoints on the mathematical formulations and their implications, with no clear consensus reached on the interpretations or applications of the discussed concepts.

Contextual Notes

Some mathematical steps and assumptions remain unresolved, particularly in the transition between different representations of mathematical objects and their physical interpretations.

Mehdi_
Messages
65
Reaction score
0
Originally Posted by garrett:
using the "limit" definition for the exponential. This is an exact expression for the rotation of a vector by a bivector. In three dimensions an arbitrary bivector, B, can be written as
B = \theta b
an amplitude, \theta , multiplying a unit bivector encoding the orientation, bb=-1. The exponential can then be written using Joe's expression for exponentiating a bivector:
U = e^{\frac{1}{2} B} = \cos(\frac{1}{2} \theta) + b \sin(\frac{1}{2} \theta)

U = e^{\frac{1}{2} B} = \cos(\frac{1}{2} \theta) + b \sin(\frac{1}{2} \theta)
we can then write:
U = e^{\frac{1}{2} \theta b} = \cos(\frac{1}{2} \theta) + b \sin(\frac{1}{2} \theta)
And if we rely on Joe's expression, r=\frac{\theta}{2} (rotor angle is always half the rotation):
U = e^{br} = \cos(r) + b \sin(r)
and
\begin{align*}U=e^T & = I \left(1 - \frac{1}{2!}r^2 + \frac{1}{4!}r^4 - \frac{1}{6!}r^6 + \dots\right) + T \left(1 - \frac{1}{3!}r^2 + \frac{1}{5!}r^4 - \frac{1}{7!}r^6 + \dots \right) \\ & = I \cos(r) + \frac{1}{r} T \sin(r)\end{align*}
we can therefore see that br = T
then : b = \frac{T}{r}
We have then Joe's expression :
U = e^{T} = \cos(r) + \frac{T}{r} \sin(r)

And from bb=-1 we can deduce:

\frac{T^2}{r^2}=-I
and therefore TT =-r^2 I (result defined by the joe previously)
 
Last edited:
Physics news on Phys.org
Relation between su(2) generators and Euler's formula in complex analysis.

Euler's formula states that, for any real number x, e^{ix} = cos(x) + i sin(x)
If however x is a Pauli matrix, Euler's formula becomes :
e^{i\sigma} = cos(\sigma) + i sin(\sigma) being understood that (\sigma_i)^2=1
Pauli matrices T= i \sigma form a basis for su(2) over R.
The matrices T could also be called generators of the Lie algebra su(2).
And the Euler's formula becomes then:
e^T = cos(\sigma) + \frac{T}{\sigma} sin(\sigma)

To every Lie group, we can associate a Lie algebra, whose underlying vector space is the tangent space of G at the identity element. But how to construct a Lie group from a Lie algebra ?
Answer: The most general elements U of the group SU(2) can be obtained by exponentiating the generators of the algebra su(2).
U = e^T = I cos(\sigma) + \frac{T}{\sigma} sin(\sigma)
 
Last edited:
The permutation symbol \epsilon_{ijk} is a three-index object sometimes called the Levi-Civita symbol.
\epsilon_{ijk}=0 if i=j or j=k or i=k
\epsilon_{ijk}=+1 if (i,j,k) \in {(1,2,3),(2,3,1),(3,1,2)}
\epsilon_{ijk}=-1 if (i,j,k) \in {(1,3,2),(3,2,1),(2,1,3)}

The symbol can be defined as the scalar triple product of unit vectors in a right-handed coordinate system: \epsilon_{ijk} \equiv \vec{x}_i . (\vec{x}_j \times \vec{x}_k)

The symbol \epsilon_{ijk} can be generalized to an arbitrary number of elements, in which case the permutation symbol is (-1)^{i(p)} , where i(p) is the number of transpositions of pairs of elements that must be composed to build up the permutation p.

The permutation symbol satisfies
\delta_{ij}\epsilon_{ijk} = 0
\epsilon_{ipq}\epsilon_{jpq} = 2\delta_{ij}
\epsilon_{ijk}\epsilon_{ijk} = 6
\epsilon_{ijk}\epsilon_{pqk} = \delta_{ip}\delta_{iq}-\delta_{iq}\delta_{jp}
where \delta_{ij} is the Kronecker delta
 
Last edited:
The simplest interpretation of the Kronecker delta is as the discrete version of the delta function defined by
\delta_{ij}=0 for i and j different
\delta_{ij}=1 for i = j

In three-space, the Kronecker delta satisfies the identities
\delta_{ii} = 3
\delta_{ij}\epsilon_{ijk} = 0
\epsilon_{ipq}\epsilon_{jpq} = 2\delta_{ij}
\epsilon_{ijk}\epsilon_{pqk} = \delta_{ip}\delta_{iq}-\delta_{iq}\delta_{jp}

where Einstein summation is implicitly assumed (i,j = 1,2,3...)

Technically, the Kronecker delta is a mixed second-rank tensor defined by the relationship
\delta^{i}_{j} = \frac{\partial x_i}{\partial x_ j}
Since the coordinates x_i and x_ j are independent for i not equal to j, and therefore :

\delta^{i}_{j} = \frac{\partial x_i}{\partial x_ k} \frac{\partial x_l}{\partial x_ j} \delta^{k}_{l}

The n \times n identity matrix I can be written in terms of the Kronecker delta as simply the matrix of the delta, I_{ij} = \delta_{ij} , or simply I = (\delta_{ij}).

The generalized Kronecker is defined however by:
\delta^{jk}_{ab}=\epsilon_{abi}\epsilon^{jki} = \delta^{j}_{a}\delta^{k}_{b}-\delta^{k}_{a}\delta^{j}_{b}
\delta_{abjk}= g_{aj}g_{bk}-g_{ak}g_{bj}
\epsilon_{aij}\epsilon^{bij} = \delta^{bi}_{ai} = 2 \delta^{b}_{a}

Actually, the generalized Kronecker delta could also be written as as a determinant :

\delta^{i_1...i_k}_{j_1...j_k} = \sum & \mbox{ sign } & \tau \delta^{i_\tau_1)}_{j_1}...\delta^{i_\tau_k}_{j_k}

= & \mbox{ det} & \left[\begin{array}{ccc}\delta^{i_1}_{j_1} & \mbox{ ... } & \delta^{i_k}_{j_1} \\ \mbox{ ... } & \mbox{ ... } & \mbox{ ... } \\ \delta^{i_1}_{j_k} & \mbox{ ... } & \delta^{i_k}_{j_k} \end{array}\right]
 
Last edited:
Under summation convention, \delta_{ij} \mbox{ A_j } = \mbox{ A_i }

The cross product C = A \times B can be written : C_i = \epsilon_{ijk}A_j B_k

The curl of A is : (\nabla \times A)_i = \epsilon_{ijk} \frac{\partial A_k}{\partial x_ j}

Orthonormality property : e_i . e_j = \delta_{ij}
 
Last edited:
Different kinds of Derivatives (extracted from Wikipedia, the free encyclopedia)

The derivative is often defined as the instantaneous rate of change of a function.
The simplest type of derivative is the derivative of a real-valued function of a single real variable; the derivative gives then the slope of the tangent to the graph of the function at a point or provides a mathematical formulation of rate of change.

A partial derivative of a function of several variables is its derivative with respect to one of the variables with the others held constant.

The total derivative however, all variables are allowed to vary.

For real valued functions from R^n to R, the total derivative is often called the gradient. An intuitive interpretation of the gradient is that it points "up": in other words, it points in the direction of fastest increase of the function. It can be used to calculate directional derivatives of scalar functions or normal directions.

Several linear combinations of partial derivatives are especially useful in the context of differential equations defined by a vector valued function R^n \rightarrow R^n. The divergence gives a measure of how much "source" or "sink" near a point there is. It can be used to calculate flux by divergence theorem. The curl measures how much "rotation" a vector field has near a point.

The other forms of derivatives will be studied individually and more extensively : Directional derivatives, Lie derivative, lie brackets, Lie bracket, Exterior derivative, Covariant derivative, Jacobian matrix, pushforward.

There is ever more other forms of derivatives that will not be studied here : Fréchet derivative, Gâteaux derivative, Exterior covariant derivative, Radon-Nikodym derivative, Kähler differential...
 
Last edited:
Chain Rule & Derivative


\frac{d}{dx} (u \mbox{ o } v) = \frac{dv}{dx}( \frac{du}{dx} \mbox{ o } v)

D( g \mbox{ o } f)(x) = D(g)(f(x)) \mbox{ o } D(f)(x)

\frac{du}{dx} = \frac{\frac{du}{dv}}{\frac{dx}{dv}} = \frac{du}{dv} \frac{dv}{dx}
 
Last edited:
Derivative of the Exponential and Logarithmic functions

\frac{d}{dx} (ln(u)) = \frac{1}{u} \frac{du}{dx}

Because log_a(x) = \frac{ln(x)}{ln(a)} we have however \frac{d}{dx} (log_a(u)) = \frac{1}{\mbox{ ln(a) u }} \frac{du}{dx}

\frac{d}{dx} (e^u) = e^u \frac{du}{dx}

\frac{d}{dx} (a^u) = ln (a)a^u \frac{du}{dx}

\frac{d}{dx} (u^v) = \frac{d}{dx} (e^{\mbox{v ln(u)}}) = e^{\mbox{v ln(u)} }} \frac{du}{dx} (\mbox{v ln(u)}) = v u^{v-1} \frac{du}{dx} \mbox{ + } u^v ln(u) \frac{dv}{dx}
 
Last edited:
The derivative at a point could be seen as a linear approximation of a function at that point.

For example, if for a given differentiable function f of one real variable, Taylor's theorem near the point a is :

f(x) = f(a) + \frac{f^{(1)}(a)}{1!} (x-a) + \frac{f^{(2)}(a)}{2!} (x-a)^2 + \frac{f^{(3)}(a)}{3!} (x-a)^3 +...+ \frac{f^{(n)}(a)}{n!} (x-a)^n + R_n

When n=1, Taylor's theorem simply becomes :

f(x) = f(a) + \frac{f^{(1)}(a)}{1!} (x-a) + R_n

The linear approximation is then obtained by dropping the remainder:

f(x) \approx f(a) + f^{(1)}(a) (x-a)

This process could therefore also be called the tangent line approximation.
The function f is then approximated by a tangent line, fact which remaind us that in differential geometry, one can attach tangent vectors to every point p of a differentiable manifold.

We can also use linear approximations for vector functions of vector variables, in which case f^{(1)}(a) is the Jacobian matrix J_f(a). The approximation is then the equation of a tangent line, plane, or hyperplane...

f(x) \approx f(a) + J_f(a) \mbox{.} (x-a)


In the more general case of Banach spaces, one has

f(x) \approx f(a) + Df(a) (x-a)

where Df(a) is the Fréchet derivative of f at a.
 
Last edited:
  • #10
Fréchet & Gâteaux derivative (mostly from Wikipedia, the free encyclopedia)

A Fréchet derivative is a derivative defined on Banach spaces.

If a function f is Fréchet differentiable at a point a, then its Fréchet derivative is :

Df(a) : R^n \rightarrow R^m \mbox{ with } Df(a)(v) = J_f(a) v

where J_f(a) denotes the Jacobian matrix of f at a

Furthermore, the partial derivatives of f are given by :

\frac{\partial f}{\partial x_i} (a) = Df(a)(e_i) = J_f(a) e_i

where e_i are the canonical basis of R^n.

Since Fréchet derivative is a linear function, the directional derivative of the function f along vector h is given by :

Df(a)(h) = \sum h_i \frac{\partial f}{\partial x_i} (a)

That bring us naturally to Gâteaux derivative which is a generalisation of the concept of directional derivative (see functional derivatives for more details).
A Gâteaux derivative could sometimes be a Fréchet differentive but unlike the other forms of derivatives, the Gâteaux derivative is not linear.
 
Last edited:
  • #11
Jacobian matrix

The Jacobian matrix is a matrix which elements are first-order partial derivatives of a vector-valued function.
It represents a linear approximation to a differentiable function near a given point.

Suppose F \mbox{ : } R^n \rightarrow R^m is a function.
Jacobian matrix of F is :

J_F(x_1, ... , x_n) = \frac{\partial (y_1, ... , y_m)}{\partial (x_1, ... , x_n)} = \left[\begin{array}{ccc}\frac{\partial (y_1)}{\partial (x_1)} & \mbox{ ... } & \frac{\partial (y_1)}{\partial (x_n)} \\ \mbox{ ... } & \mbox{ ... } & \mbox{ ... } \\ \frac{\partial (y_m)}{\partial (x_1)}& \mbox{ ... } & \frac{\partial (y_m)}{\partial (x_n)}\end{array}\right]

The determinant of J_F is the Jacobian determinant \mid J_F \mid

| J_F(x_1, ... , x_n) | = det ( \frac{\partial (y_1, ... , y_m)}{\partial (x_1, ... , x_n) ) = \left | \begin{array}{ccc}\frac{\partial (y_1)}{\partial (x_1)} & \mbox{ ... } & \frac{\partial (y_1)}{\partial (x_n)} \\ \mbox{ ... } & \mbox{ ... } & \mbox{ ... } \\ \frac{\partial (y_m)}{\partial (x_1)}& \mbox{ ... } & \frac{\partial (y_m)}{\partial (x_n)}\end{array}\right |
 
Last edited:
  • #12
Examples involving the Jacobian determinant \mid J_F \mid

We'll begin by looking first at an example which show how a definite integral is affected by a change of variables.
Suppose we want to evaluate the definite integral \int^{a}_{b} f(u) du

\int^{a}_{b} f(u) du = \int^{c}_{d} f(u(x)) \frac{du}{dx}dx

We see that the endpoints are changed and there is a new factor \frac{du}{dx}.
It can also be written du=\frac{du}{dx}dx

The new factor \frac{du}{dx} is a partial derivative which can then be considered as a 1 \times 1 Jacobian matrix J_{u(x)}.

Often, because the limits of integration are not easily interchangeable one makes a change of variables to rewrite the integral on a different region of integration. To do that, the function must be changed to the new coordinates (i.e. Passage from cartesian to polar coordinates).

As an example, let's consider a domaine D = (x^2+y^2 \leq 9, x^2+y^2 \geq 4, y \geq 0) , that is the circular crown in the semiplane of positive y. Therefore the transformed domain will be the following rectangle:
T = (2 \geq \rho \leq 3, 0 \geq \phi \leq \pi)

The Jacobian determinant of that transformation is the following:

| J_{f(x,y)} | = det | \frac{\partial (x,y)}{\partial (\rho, \phi) }| = \left | \begin{array}{ccc} cos(\phi) & - \rho sin(\phi) \\ sin(\phi) & \rho cos(\phi)\end{array}\right | = \rho

which has been got by inserting the partial derivatives of x = \rho cos(\phi) and y = \rho sin(\phi)

It's then possible to define the integral for the change of variables in polar coordinates:

\int \int_{D} f(x,y) dx dy = \int \int_{T} f(\rho cos(\phi),\rho sin(\phi)) \mbox{ det( J_{f(x,y)} ) } d \rho d \phi) = \int \int_{T} f(\rho cos(\phi),\rho sin(\phi)) \mbox{ \rho } d \rho d \phi)
 
Last edited:
  • #13
Differential forms

A differential 0-form on R^3 is a scalar function w \mbox{ : } D \rightarrow R^3 of class C^1 on a domain D in R^3.

A differential 1-form on R^3 is an expression of the form

w = Adx + Bdy + Cdz

where F(A(x,y,z),B(x,y,z),C(x,y,z)) \mbox{ : } D \rightarrow R^3 is a vector field on a domain D in R^3 for which the functions A,B,C belong to class C^1.
F(A(x,y,z),B(x,y,z),C(x,y,z)) could be also written F(a_i) if (a_1)=A(x_i), (a_2)=B(x_i),....

The differential 1-form could then be written w = \sum a_i dx_i or simply w = a_i dx^i


A differential 2-form on R^3 is an expression of the form

w = Ady \wedge dz + Bdz \wedge dx + Cdx \wedge dy

where F(A(x,y,z),B(x,y,z),C(x,y,z)) \mbox{ : } D \rightarrow R^3 is a vector field on a domain D in R^3 for which the functions A,B,C belong to class C^1.

If \alpha and \beta two 1-forms:

\alpha = f_1 dx^1 + f_2 dx^2 and \beta = g_1 dx^1 + g_2 dx^2

\alpha \wedge \beta = (f_1 dx^1 + f_2 dx^2 ) \wedge (g_1 dx^1 + g_2 dx^2)
= f_1g_2 dx^1 \wedge dx^2 \mbox{+} f_2g_1 dx^2\wedge dx^1
= f_1g_2 dx^1 \wedge dx^2 \mbox{-} f_2g_1 dx^1\wedge dx^2
= (f_1g_2 \mbox{ - } f_2g_1) dx^1\wedge dx^2
= \left | \begin{array}{ccc} f_1 & f_2 \\ g_1 & g_2 \end{array}\right | \mbox{ } dx^1\wedge dx^2
Then

\alpha \wedge \beta = \sum_{i < j} (f_i g_j \mbox{ - } f_j g_i) dx^i \wedge dx^j


A differential k-form could be written

\xi = \sum a_{i_1 ... i_k} (x) dx^{i_1} \wedge ... \wedge dx^{i_k}

\gamma = \sum a_{i_1 ... i_j} (x) dx^{i_1} \wedge ... \wedge dx^{i_j}

\xi \wedge \gamma = \sum a_{i_1 ... i_k} (x) b_{i_1 ... i_j} (x) dx^{i_1} \wedge ... \wedge dx^{i_{k+j}}

which is differential (k+j)-form.

The coefficients of a differential n-form change under a change of basis by multiplication by the Jacobian.
If w is a differential n-form

\omega = \omega (x) dx^1 \wedge ... \wedge dx^n

then if \bar{x^i} are new coordinates for x, then, in these new coordinates

\omega = \omega (x) J \mbox{ } d \bar{x^1} \wedge ... \wedge d \bar{x^n}

where J = \frac{ \partial x^i }{ \partial \bar{x^j} }
 
Last edited:
  • #14
Now if F and G are a 0-form, \alpha and \beta could be written:

\alpha = dF = \frac{\partial F}{\partial x^1} dx^1 + \frac{\partial F}{\partial x^2} dx^2
and
\beta = dG = \frac{\partial G}{\partial x^1} dx^1 + \frac{\partial G}{\partial x^2} dx^2

dF \wedge dG = (\frac{\partial F}{\partial x^1} dx^1 + \frac{\partial F}{\partial x^2} dx^2) \wedge (\frac{\partial G}{\partial x^1} dx^1 + \frac{\partial G}{\partial x^2} dx^2 )

= (\frac{\partial F}{\partial x^1} \frac{\partial G}{\partial x^2} ) dx^1 \wedge dx^2 \mbox{ - } (\frac{\partial F}{\partial x^2} \frac{\partial G}{\partial x^1} ) dx^1 \wedge dx^2

= ( \frac{\partial F}{\partial x^1} \frac{\partial G}{\partial x^2} \mbox{ - } \frac{\partial F}{\partial x^2} \frac{\partial G}{\partial x^1} ) dx^1 \wedge dx^2

= | \frac{\partial (F,G)}{\partial (x^1,x^2) }| dx^1 \wedge dx^2

The 2-form dF \wedge dG has then be converted to dx^1dx^2-coordinates.

dF \wedge dG = | \frac{\partial (F,G)}{\partial (x^1,x^2) }| dx^1 \wedge dx^2
 
Last edited:
  • #15
The exterior derivative of a k-differential form

d \xi = \sum_{i_1 < i_k} d a_{i_1 ... i_k} (x) dx^{i_1} \wedge ... \wedge dx^{i_k}

d \xi = \sum_{i_1 < i_k} d a_{i_1 ... i_k} (x) \wedge dx^{i_1} \wedge ... \wedge dx^{i_k}

example

w = \sum_i f_i dx_i

w = b_1(x_1,x_2) dx_1 + b_2(x_1,x_2) dx_2

dw = db_1(x_1,x_2) dx_1 + db_2(x_1,x_2) dx_2

dw = db_1(x_1,x_2) \wedge dx_1 + db_2(x_1,x_2) \wedge dx_2

dw = (\frac{\partial b_1}{\partial x^1} dx_1 + \frac{\partial b_1}{\partial x^2} dx_2) \wedge dx_1 + (\frac{\partial b_2}{\partial x^1} dx_1 + \frac{\partial b_2}{\partial x^2} dx_2) \wedge dx_2
 
Last edited:
  • #16
dw = (\frac{\partial b_1}{\partial x^1} dx_1 + \frac{\partial b_1}{\partial x^2} dx_2) \wedge dx_1 + (\frac{\partial b_2}{\partial x^1} dx_1 + \frac{\partial b_2}{\partial x^2} dx_2) \wedge dx_2

dw = (\frac{\partial b_1}{\partial x^1} dx_1 \wedge dx_1 + \frac{\partial b_1}{\partial x^2} dx_2 \wedge dx_1) + (\frac{\partial b_2}{\partial x^1} dx_1 \wedge dx_2 + \frac{\partial b_2}{\partial x^2} dx_2 \wedge dx_2 )

dw = \frac{\partial b_1}{\partial x^2} dx_2 \wedge dx_1 + \frac{\partial b_2}{\partial x^1} dx_1 \wedge dx_2

dw = - \frac{\partial b_1}{\partial x^2} dx_1 \wedge dx_2 + \frac{\partial b_2}{\partial x^1} dx_1 \wedge dx_2

dw = \frac{\partial b_2}{\partial x^1} dx_1 \wedge dx_2 - \frac{\partial b_1}{\partial x^2} dx_1 \wedge dx_2

dw = ( \frac{\partial b_2}{\partial x^1} - \frac{\partial b_1}{\partial x^2} ) dx_1 \wedge dx_2
 
Last edited:
  • #17
A differential k-form should be written

\omega = \sum_{i_1 < i_k} a_{i_1 ... i_k} (x^m) dx^{i_1} \wedge ... \wedge dx^{i_k}

The term i_1 < i_k is very important.

It could however also be written without \sum_{i_1 < i_k}.
The new representation use the convenience of the summation convention.

\omega = \frac{1}{k!} a_{i_1 ... i_k} (x^m) dx^{i_1} \wedge ... \wedge dx^{i_k}

We can see that the term \frac{1}{k!} arises in this alternative way of writing a differential k-form in order to show clearly that we do not repeat k! times each term a_{i_1 ... i_k} (x^m).
 
Last edited:
  • #18
Exterior algebra is the algebra of the exterior product \wedge, also called an alternating algebra or Grassmann algebra.

with the properties (if \omega and \gamma differential forms )

\omega \wedge \omega = 0

\omega \wedge \gamma = \mbox{ - } \gamma \wedge \omega

Exterior algebra of a given vector space V over a field K is denoted by Λ(V) or Λ*(V).

The exterior algebra can be written as the direct sum of each of the k-th powers:

\bigwedge (V) = \bigoplus^{ n }_{k=0} \bigwedge^{k} V

Therefore

\bigwedge (V) = \bigwedge^{0} V \bigoplus \bigwedge^{1} V \bigoplus ... \bigoplus \bigwedge^{k} V

where \bigwedge^{0} V = k and \bigwedge^{1} V = V

The dimension of \bigwedge^{k} V is n choose k, \left(\begin{array}{cc} n \\ k \end{array}\right)

The dimension of \bigwedge (V) is then equal to the sum of the binomial coefficients, \sum^{n}_{k=0} \left(\begin{array}{cc} n \\ k \end{array}\right), which is 2^n
 
Last edited:
  • #19
If E is a vector space, E* is then it's dual space.

\bigwedge^{r} E* is the vector space of multilinear alternating r-forms on E.

We have \bigwedge^{0} E* = R and \bigwedge^{1} E* = E*

\bigwedge^{1} E*, the space of differential 1-forms, coincides with the dual space T*(E) which is the cotangent space.

The elements of \bigwedge^{1} E*, in terms of natural basis dx^i, have the representation:

\omega = \omega_ i (x^j) dx^i which is a 1-form.

The elements of \bigwedge^{0} E* is referred to as the space of the forms of degree zero which is the space of the functions f(x).
 
Last edited:
  • #20
Mehdi_ said:
To every Lie group, we can associate a Lie algebra, whose underlying vector space is the tangent space of G at the identity element. But how to construct a Lie group from a Lie algebra ?
Answer: The most general elements U of the group SU(2) can be obtained by exponentiating the generators of the algebra su(2).
U = e^T = I cos(\sigma) + \frac{T}{\sigma} sin(\sigma)

I'm sure you realize that's not an answer to your question. For your question "But how to construct a Lie group from a Lie algebra ?", the answer is: generally you can't. It's just for simply connected group manifolds that you can apply the method suggested for SU(2).

Daniel.
 
  • #21
Hi dextercioby, welcome to "From Simple Groups to Quantum Field Theory " thread.
 
  • #22
The set of tangent vectors at a point p forms a vector space called the tangent space at p T_{p}.

If p is a point in an n-dimensional compact manifold M, the tangent space of M at p is then denoted T_{p} M

The collection of tangent spaces T_{p_{ i }} M on a manifold M forms a vector bundle called the tangent bundle T M.

T M = \bigcup_{i=0}^{\infty} T_{p_i} M

The tangent bundle is a special class a vector bundle which means that it is also a special class of fiber bundle.

A fiber of a map f : X \longrightarrow Y is the preimage of an element y \in Y. That is,

f^{-1} (y) = [ x \in X : f(x) = y ].

For instance, when f(z) = z^2, every fiber consists of two points [ -z , z ], except for the fiber over 0, which has one point, the Ker (f) .

If \gamma however is a smooth curve passing through p, then the derivative of \gamma at p is a vector in the tangent space of M at p, T_{p} M.

A vector field is an assignment of a tangent vector for each point p of a manifold M.

The collection of tangent vectors forms the tangent bundle, and a vector field v is a section of this bundle T M.

A vector field on M is then a map v which assigns to each point p \in M a tangent vector v(p).

v(p) = v_p \in T_p M

A tangent vector is the manifold version of a directional derivative at a point.

A vector field v acts on a function f by the directional derivative d_v on the function

d_v (f) = v . d(f) \equiv \bigtriangledown_v f
 
Last edited:
  • #23
Scalar fields

Scalar field is a map over some space of scalar values. It is a map of values with no direction.

A simple example of a scalar field is a map of the temperature distribution in a room.

A scalar field could be viewed as a map f : R^n \longrightarrow R which assigns each point x in a n-dimensional space V with a scalar function f(x_i).

The position vector of a point x, could be written in the form

\vec{ x } \equiv ( x_1, x_2, x_3 ) \equiv x_1 \vec{ e_1 } + x_2 \vec{ e_2 } + x_3 \vec{ e_3 }

If with each point x at the position ( x_1, x_2, x_3 ) there corresponds a scalar f ( x_1, x_2, x_3 ) such that

( x_1, x_2, x_3 ) \longrightarrow f ( x_1, x_2, x_3 )

then the values of f ( x_1, x_2, x_3 ) associated with all the points in V define a scalar field over V.

Example 1: f ( x_1, x_2, x_3 ) = 9 (x_1)^4 + 6 x_2x_ 3 - (x_3)^5 + 7

Example 2: f ( x ) = 9 x^4 + 6 x^2 - 5
 
Last edited:
  • #24
Vector fields, vector valued functions and parametric equations

A vector field could be viewed as a map f : R^n \longrightarrow R^n that assigns each point x in a n-dimensional space V with a vector valued function f ( x_i(t) ) whose range is also n-dimensional.

A vector valued function r(t) is a function where the domain is a subset of the real numbers and the range is a vector : r(t) = r_1(t) \vec{ e_1 } + r_2(t) \vec{ e_2 } + r_3(t) \vec{ e_3 }
Vector valued functions can also be referred to in a different notation : r(t) = <r_1(t), r_2(t), r_3(t) >

Actually, there is an equivalence between vector valued functions and parametric equations.

As an example let's consider the vector valued function

r(t) = 2 . cos ( t ) . \vec{ i } + 2 . sin ( t ) . \vec{ j } + \frac{1}{2} . t . \vec{ k }.

To understand this function, consider the parametric curve

x = 2 . cos ( t )
y = 2 . sin ( t )
z = \frac{1}{2} t

The two equations x = 2 . cos ( t ) and y = 2 . sin ( t ) describe a point in the xy-plane that is moving in a circle (This is because x^2 + y^2 = 4 ).

Meanwhile, the value of z = \frac{1}{2} t increases at t increases.

The result is a helix, a spiral curve that wraps around the cylinder x^2 + y^2 = 4 .

Example of vector field : let's sketch first the following direction field.

\vec{ F } ( x, y ) = - y \vec{ i } + x \vec{ j }

To graph the vector field we need to get some values of the function.
This means plugging in some points into the vector valued function.

\vec{ F } ( \frac{1}{2}, \frac{1}{2} ) = - \frac{1}{2} \vec{ i } + \frac{1}{2} \vec{ j }

\vec{ F } ( \frac{1}{2}, - \frac{1}{2} ) = \frac{1}{2} \vec{ i } + \frac{1}{2} \vec{ j }

\vec{ F } ( \frac{3}{2}, \frac{1}{4} ) = - \frac{1}{4} \vec{ i } + \frac{3}{2} \vec{ j }

At the point ( \frac{1}{2}, \frac{1}{2} ) we will plot the vector - \frac{1}{2} \vec{ i } + \frac{1}{2} \vec{ j }.

Likewise, the third evaluation tells us that at the point ( \frac{3}{2}, \frac{1}{4} ) we will plot the vector - \frac{1}{4} \vec{ i } + \frac{3}{2} \vec{ j }


We can continue in this fashion plotting vectors for several points and we’ll get the sketch of the vector field.
 
Last edited:
  • #25
Vector fields

Tangent vector fields are defined on manifolds as sections of the manifold's tangent bundle.

But more generally, a vector fields on a manifold could be simply defined as sections of the vector bundle (kind of fiber bundle).

Let X be a vector field on V.

X = \sum^{n}_{i=1} X (x_i) \frac{\partial }{\partial x_i}

Each X (x_i) is by definition just a differentiable function on V.

The tangent space basis \frac{\partial }{\partial x_i} are also the vector field basis.
They are isomorphic to the euclidean basis \vec{ e_i }.

\frac{\partial }{\partial x_i} \equiv \vec{ e_i } \equiv \vec{ i }, \vec{ j } \mbox{ ou } \vec{ k }

The vector field \vec{r}(t) = 2 . cos ( t ) . \vec{ i } + 2 . sin ( t ) . \vec{ j } + \frac{1}{2} . t . \vec{ k } could then also be written :

\vec{r}(t) = 2 . cos ( t ) \frac{\partial }{\partial x_1} + 2 . sin ( t ) \frac{\partial }{\partial x_2} + \frac{1}{2} . t \frac{\partial }{\partial x_3}
 
Last edited:
  • #26
Gradient and directional derivatives

Gradient is commonly used to describe the measure of the slope (derivative) of a function.

For vector-valued function, the gradient is then the Jacobian.

The gradient of a scalar field is a vector field which points in the direction of the greatest rate of increase of the scalar field, and whose magnitude is the greatest rate of change.

The gradient of a function f(x) could be denoted by grad(f) or equivalent by \nabla f where the symbol \nabla is variously known as Nabla or Del

grad(f) = \nabla f = < f_x, f_y> where f_x and f_y are partial derivatives

For example, the gradient of f (x,y,z) = 2x + {3y}^2 - sin(z) is the vector

\nabla f = ( \frac{ \partial f }{ \partial x^1} + \frac{ \partial f }{ \partial x^2} + \frac{ \partial f }{ \partial x^3} )^T = ( 2, 6y, - cos(z) )^T

The directional derivative (in terms of the gradient) D_{\vec{v}} f of a scalar function f( \vec{x} ) = f(x_i) along a vector \vec{v} = (v_1 ... v_n)^T is the function

D_{\vec{v}} f = \nabla f . \vec{v}

where the dot denotes the dot product (Euclidean inner product) , \nabla f the gradient of the function f and \vec{v} a unit vector

Therefore

D_{\vec{w}} f = \nabla f . \frac{ \vec{w} }{ \vec{| w |} }


\frac{ \vec{w} }{ \vec{| w |} } = <cos (\theta), sin(\theta) >

example : f (x,y) = x^2 + y^2 and \vec{v} = <3, 4>

The directional derivative is

\frac{ \vec{v} }{ \vec{| v |} } = \frac{ 1 }{ \sqrt{ 9 + 16} } } <3, 4> = < \frac{3} {5} , \frac{4} {5} >

f_x = 2 x and f_y = 2 y

D_{\vec{v}} f (x,y) = (2 x ) \frac{3} {5} + (2 y ) \frac{4} {5} = \frac{ 6 x + 8 y } {5}

At the point (1,2,5)

D_{\vec{v}} f (1,2) = \frac{ 22 } {5}

The directional derivative in a general direction is then

D_{\vec{v}} f = \frac{ d f }{ ds } = \frac{\partial f}{\partial x_1} \frac{d x_1}{ ds } + \frac{\partial f}{\partial x_2} \frac{ d x_2}{ ds }+ \frac{\partial f}{\partial x_3} \frac{ d x_3}{ ds } = f_{x_1} \frac{ d x_1}{ ds } + f_{x_2} \frac{ d x_2}{ ds }+ f_{x_3} \frac{ d x_3}{ ds }

If \frac{ \vec{v} }{ \vec{| v |} } = < \frac{ d x_1}{ ds } , \frac{ d x_2}{ ds } , \frac{ d x_3}{ ds } >

d s is called element of arc or element of the curve C and s is called arc length of the curve C.
\vec{v} is a unit vector tangent to the curve C and directed in the direction of growing s

Two points of the curve C at the positions s and (s+h), determine a chord whose direction is given by the vector x(x+h)-x(s)
The vector
\vec{v} Gradient and directional derivatives <br /> <br /> Gradient is commonly used to describe the measure of the slope (derivative) of a function. <br /> <br /> For vector-valued function, the gradient is then the Jacobian.<br /> <br /> The gradient of a scalar field is a vector field which points in the direction of the greatest rate of increase of the scalar field, and whose magnitude is the greatest rate of change.<br /> <br /> The gradient of a function f(x) could be denoted by grad(f) or equivalent by \nabla f where the symbol \nabla is variously known as Nabla or Del<br /> <br /> grad(f) = \nabla f = &amp;lt; f_x, f_y&amp;gt; where f_x and f_y are partial derivatives <br /> <br /> For example, the gradient of f (x,y,z) = 2x + {3y}^2 - sin(z) is the vector <br /> <br /> \nabla f = ( \frac{ \partial f }{ \partial x^1} + \frac{ \partial f }{ \partial x^2} + \frac{ \partial f }{ \partial x^3} )^T = ( 2, 6y, - cos(z) )^T<br /> <br /> The directional derivative (in terms of the gradient) D_{\vec{v}} f of a scalar function f( \vec{x} ) = f(x_i) along a vector \vec{v} = (v_1 ... v_n)^T is the function <br /> <br /> D_{\vec{v}} f = \nabla f . \vec{v}<br /> <br /> where the dot denotes the dot product (Euclidean inner product) , \nabla f the gradient of the function f and \vec{v} a unit vector<br /> <br /> Therefore<br /> <br /> D_{\vec{w}} f = \nabla f . \frac{ \vec{w} }{ \vec{| w |} }<br /> <br /> <br /> \frac{ \vec{w} }{ \vec{| w |} } = &amp;lt;cos (\theta), sin(\theta) &amp;gt;<br /> <br /> example : f (x,y) = x^2 + y^2 and \vec{v} = &amp;lt;3, 4&amp;gt; <br /> <br /> The directional derivative is <br /> <br /> \frac{ \vec{v} }{ \vec{| v |} } = \frac{ 1 }{ \sqrt{ 9 + 16} } } &amp;lt;3, 4&amp;gt; = &amp;lt; \frac{3} {5} , \frac{4} {5} &amp;gt;<br /> <br /> f_x = 2 x and f_y = 2 y <br /> <br /> D_{\vec{v}} f (x,y) = (2 x ) \frac{3} {5} + (2 y ) \frac{4} {5} = \frac{ 6 x + 8 y } {5}<br /> <br /> At the point (1,2,5)<br /> <br /> D_{\vec{v}} f (1,2) = \frac{ 22 } {5}<br /> <br /> The directional derivative in a general direction is then <br /> <br /> D_{\vec{v}} f = \frac{ d f }{ ds } = \frac{\partial f}{\partial x_1} \frac{d x_1}{ ds } + \frac{\partial f}{\partial x_2} \frac{ d x_2}{ ds }+ \frac{\partial f}{\partial x_3} \frac{ d x_3}{ ds } = f_{x_1} \frac{ d x_1}{ ds } + f_{x_2} \frac{ d x_2}{ ds }+ f_{x_3} \frac{ d x_3}{ ds }<br /> <br /> If \frac{ \vec{v} }{ \vec{| v |} } = &amp;lt; \frac{ d x_1}{ ds } , \frac{ d x_2}{ ds } , \frac{ d x_3}{ ds } &amp;gt;<br /> <br /> d s is called element of arc or element of the curve C and s is called arc length of the curve C.<br /> \vec{v} is a unit vector tangent to the curve C and directed in the direction of growing s<br /> <br /> Two points of the curve C at the positions s and (s+h), determine a chord whose direction is given by the vector x(x+h)-x(s) <br /> The vector <br /> \vec{v} = \liminf_0
 
Last edited:
  • #27
Gradient and directional derivatives

Gradient is commonly used to describe the measure of the slope (derivative) of a function.

For vector-valued function, the gradient is then the Jacobian.

The gradient of a scalar field is a vector field which points in the direction of the greatest rate of increase of the scalar field, and whose magnitude is the greatest rate of change.

The gradient of a function f(x) could be denoted by grad(f) or equivalent by \nabla f where the symbol \nabla is variously known as Nabla or Del

grad(f) = \nabla f = &lt; f_x, f_y&gt; where f_x and f_y are partial derivatives

For example, the gradient of f (x,y,z) = 2x + {3y}^2 - sin(z) is the vector

\nabla f = ( \frac{ \partial f }{ \partial x^1} + \frac{ \partial f }{ \partial x^2} + \frac{ \partial f }{ \partial x^3} )^T = ( 2, 6y, - cos(z) )^T

The directional derivative (in terms of the gradient) D_{\vec{v}} f of a scalar function f( \vec{x} ) = f(x_i) along a vector \vec{v} = (v_1 ... v_n)^T is the function

D_{\vec{v}} f = \nabla f . \vec{v}

where the dot denotes the dot product (Euclidean inner product) , \nabla f the gradient of the function f and \vec{v} a unit vector

Therefore

D_{\vec{w}} f = \nabla f . \frac{ \vec{w} }{ \vec{| w |} }


\frac{ \vec{w} }{ \vec{| w |} } = &lt;cos (\theta), sin(\theta) &gt;

example : f (x,y) = x^2 + y^2 and \vec{v} = &lt;3, 4&gt;

The directional derivative is

\frac{ \vec{v} }{ \vec{| v |} } = \frac{ 1 }{ \sqrt{ 9 + 16} } } &lt;3, 4&gt; = &lt; \frac{3} {5} , \frac{4} {5} &gt;

f_x = 2 x and f_y = 2 y

D_{\vec{v}} f (x,y) = (2 x ) \frac{3} {5} + (2 y ) \frac{4} {5} = \frac{ 6 x + 8 y } {5}

At the point (1,2,5)

D_{\vec{v}} f (1,2) = \frac{ 22 } {5}

The directional derivative in a general direction is then

D_{\vec{v}} f = \frac{ d f }{ ds } = \frac{\partial f}{\partial x_1} \frac{d x_1}{ ds } + \frac{\partial f}{\partial x_2} \frac{ d x_2}{ ds }+ \frac{\partial f}{\partial x_3} \frac{ d x_3}{ ds } = f_{x_1} \frac{ d x_1}{ ds } + f_{x_2} \frac{ d x_2}{ ds }+ f_{x_3} \frac{ d x_3}{ ds }

If \frac{ \vec{v} }{ \vec{| v |} } = &lt; \frac{ d x_1}{ ds } , \frac{ d x_2}{ ds } , \frac{ d x_3}{ ds } &gt;

d s is called element of arc or element of the curve C and s is called arc length of the curve C.
\vec{v} is a unit vector tangent to the curve C and directed in the direction of growing s

Two points of the curve C at the positions s and (s+h), determine a chord whose direction is given by the vector x(s+h)-x(s)

\vec{v} = lim \frac{ x(s+h)-x(s)} {h} = \frac{ dx } {ds}

The vector \vec{v} is then called the unit tangent vector to the curve C at the point x(s).

This vector is a unit vector because

| \vec{v}|^2 = \vec{v} . \vec{v} = \frac{ dx } {ds} . \frac{ dx } {ds} = 1
 
Last edited:
  • #28
Question : How to erase post #26 ?
 
  • #29
Divergence

The divergence of a vector field F, denoted div (F) or \nabla.F, is defined by a limit of the surface integral

$ \nabla.F \equiv \displaystyle{\lim_{V \rightarrow 0}} \frac{ \oint_{ \partial V } F.da}{ V } $

where the surface integral gives the value of F integrated over a closed infinitesimal boundary surface \partial V surrounding a volume element V, which is taken to size zero using a limiting process.

A closed infinitesimal boundary surface could be viewed as a sphere of radius 0.

Therefore the divergence could also be interpreted as an operator that measures a vector field's tendency to originate from or converge upon a given point.

The divergence of a continuously differentiable vector field F = F_x i + F_y j + F_z k \equiv (F_x, F_y, F_z) is defined to be the scalar-valued function:

div (F) = \nabla.F = \frac{ \partial F_x }{ \partial x} + \frac{ \partial F_y }{ \partial y} + \frac{ \partial F_z }{ \partial z}

The divergence of a three dimensional vector field is the extent to which the vector field flow behaves like a source or a sink at a given point.

An alternative equivalent definition, gives the divergence as the derivative of the net flow of the vector field across the surface of a small sphere relative to the volume of the sphere.

In physical terms, the divergence of a vector field is the rate at which "density" exits a given region of space. The definition of the divergence therefore follows naturally by noting that, in the absence of the creation or destruction of matter, the density within a region of space can change only by having it flow into or out of the region. By measuring the net flux of content passing through a surface surrounding the region of space, it is therefore immediately possible to say how the density of the interior has changed. This property is fundamental in physics, where it goes by the name "principle of continuity." When stated as a formal theorem, it is called the divergence theorem, also known as Gauss's theorem.

The divergence theorem (Gauss' theorem, Ostrogradsky's theorem, or Ostrogradsky–Gauss theorem) is a result that relates the outward flow of a vector field on a surface to the behaviour of the vector field inside the surface.

\int_{V} (\nabla.F) dV= \oint_{ \partial V } F.da

The divergence of a linear transformation of a unit vector represented by a matrix A is given by the elegant formula

\nabla. \frac{ A \vec{x} }{ \vec{| x |} } = \frac{ Tr(A) }{ \vec{| x |} } - \frac{ \vec{x}^T Tr(A) \vec{x} }{ \vec{| x |^3} }

The concept of divergence can be generalized to tensor fields, where it is a contraction of what is known as the covariant derivative (also called the semicolon derivative), written

\nabla.A = A^{a}_{\mbox{ };b}

A^{a}_{\mbox{ };b} = \frac{ \partial A^a }{ \partial x^b } + \Gamma^{a}_{bk} A^{k} = A^{a}_{\mbox{ },b} + \Gamma^{a}_{bk} A^{k}

where \Gamma^{a}_{bk} is a Christoffel symbol
 
Last edited:
  • #30
The curl operator

The curl of a vector field F, denoted curl(F) or \nabla \times F, is defined by a limit of the surface integral below where the magnitude of \nabla \times F is the limiting value of circulation per unit area.

$ (\nabla \times F) . \hat{n} \equiv \displaystyle{\lim_{A \rightarrow 0}} \frac{ \oint_{ C } F.ds}{ A } $

where the right side is a line integral around an infinitesimal region of area A that is allowed to shrink to zero via a limiting process and \hat{n} is the unit normal vector to this region.

It can also be written as a special case of Stokes' theorem in which F is a vector field and M is an oriented, compact embedded 2-manifold with boundary in R^3, and a generalization of Green's theorem from the plane into three-dimensional space.

$ \int_{S} (\nabla \times F) . dS = \oint_{ C } F.dl $

Each differential area (\nabla \times F) . dS gives the line integral about that area since, by definition, the \nabla \times F is the circulation per unit area.

The physical significance of the curl of a vector field is the amount of "rotation" or angular momentum of the contents of given region of space. It arises in fluid mechanics and elasticity theory. It is also fundamental in the theory of electromagnetism , where it arises in two of the four Maxwell equations.

In Cartesian coordinates, the curl is defined by

curl(F) = \nabla \times F = ( \frac{ \partial F_z }{ \partial y} - \frac{ \partial F_y }{ \partial z} ) \hat{x} + ( \frac{ \partial F_x }{ \partial z} - \frac{ \partial F_z }{ \partial x} ) \hat{y} + ( \frac{ \partial F_y }{ \partial x} - \frac{ \partial F_x }{ \partial y} ) \hat{z}

which can also be written

\nabla \times F = \left | \begin{array}{ccc} \hat{x} &amp; \hat{y} &amp; \hat{z} \\ \frac{\partial}{\partial x} &amp; \frac{\partial}{\partial y} &amp; \frac{\partial}{\partial z} \\ F_x &amp; F_y &amp; F_z \end{array}\right |

A somewhat more elegant formulation of the curl is given by the matrix operator equation

\nabla \times F = \left [ \begin{array}{ccc} 0 &amp; - \frac{\partial}{\partial z} &amp; \frac{\partial}{\partial y} \\ \frac{\partial}{\partial z} &amp; 0 &amp; - \frac{\partial}{\partial x} \\ - \frac{\partial}{\partial y} &amp; \frac{\partial}{\partial x} &amp; 0 \end{array}\right ] F

The curl can be generalized from a vector field to a tensor field as

(\nabla \times F)^i = \epsilon^{ijk} F_{k ; j} = \epsilon^{ijk} \nabla_j F_k

where \epsilon^{ijk} is the permutation tensor and ";" and \nabla_j denotes the covariant derivative.
 
Last edited:

Similar threads

Replies
3
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 12 ·
Replies
12
Views
5K
  • · Replies 8 ·
Replies
8
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 9 ·
Replies
9
Views
1K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 0 ·
Replies
0
Views
2K