Electric and magnetic constants are tensors

Click For Summary

Discussion Overview

The discussion centers around the nature of tensors, specifically in relation to electric and magnetic constants. Participants explore definitions, properties, and examples of tensors, as well as the distinction between tensors and their components. The conversation includes technical explanations and conceptual clarifications.

Discussion Character

  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • Some participants define a tensor as a multilinear map from a vector space and its dual to the Reals, using the dot product as an example of a (0,2) tensor.
  • Others argue that the dot product is not a tensor but a linear function, suggesting that tensors are multilinear functions.
  • One participant explains the construction of tensors using vectors and matrices, emphasizing the importance of how components change under coordinate transformations.
  • Another participant introduces the concept of rank in tensors, noting that the rank indicates the number of indices required to specify its components.
  • There is a discussion about transformation laws for tensors, with some participants asserting that these laws are not fundamental to the definition of tensors but are rather consistency checks.
  • Participants express confusion regarding the distinction between tensors and their components, with some clarifying that notation like E_{ijk} refers to the components of a tensor, not the tensor itself.

Areas of Agreement / Disagreement

Participants express differing views on the definition and nature of tensors, particularly regarding the relationship between tensors and their components. No consensus is reached on these definitions or the implications of transformation laws.

Contextual Notes

There are unresolved aspects regarding the definitions of tensors, the implications of transformation laws, and the distinction between tensors and their components. Some statements depend on specific mathematical conventions that may vary among participants.

dervast
Messages
132
Reaction score
1
What a tensor is .? I have found a text in my book that says that the electric and magnetic constants are tensors.. Do u have something in mind?
Thx a lot
 
Physics news on Phys.org
A tensor is simply a multilinear map (a map that's linear in each variable) from a vector space and the dual of the vector space to the Reals.
A very simple example is the dot product. It takes in two 2 vectors and gives a Real number.It is linear in both varibales. Thus the dot product is a (0 2) tensor.
 
I am not sure i have completely understabd
 
Tzar said:
A tensor is simply a multilinear map (a map that's linear in each variable) from a vector space and the dual of the vector space to the Reals.
A very simple example is the dot product. It takes in two 2 vectors and gives a Real number.It is linear in both varibales. Thus the dot product is a (0 2) tensor.
No, that is not right. The dot product is not a tensor, nor is the result of a dot product a (0,2) tensor-it is a (0,0) tensor a.k.a. scalar.

dervast said:
What a tensor is .? I have found a text in my book that says that the electric and magnetic constants are tensors.. Do u have something in mind?
Thx a lot
You know that we can take several numbers and form a vector. Simmiliarly we can take N vectors of length N and produce an N by N matrix. One way we could do this is like this:
We have a vector, \vec{A} whose components in some coordinate system are A_1, A_2, A_3...A_N. When I write Ai, I mean some particular component of \vec{A}. Also we have the vector \vec{B} with components B_1, B_2, B_3...B_N. Now we can form the matrix D by saying that the element in the i'th row and the jth collumn is:
D_{ij}=A_iB_j
Similarly, I could create an object with three indices:
E_{ijk}=A_iB_jC_k
And so on. Are these tensors? Not necessarily. Notice that the components of the vectors were defined with respect to some coordinate system. We have said nothing about these components would change if we were to change the coordinate system. What makes a tensor a tensor is the way its components change when you change coordinate systems. The rank of a tensor is the number of indices required to specify its compnents.

If we are talking about rectangular coordinates then the components transform in the obvious way: First the components are written as the projections of the vector on each of the basis vectors. When you change coordinate systems you are changing basis vectors, so now you just find the new components along the new basis vectors. This transormation can be represented as a matrix. If the components of this matrix are a_{ij} and the components of \vec{A} in the new coordinate system are written as \hat{A_1}, \hat{A_2}, \hat{A_3}...\hat{A_N}, then the new coordinates are related to the old by:
\hat{A_i}=\sum_{j=1}^n a_{ij}A_j
For a tensor of rank greater than one we have:
\hat{D_{ik}}=\sum_{j=1}^n \sum_{m=1}^n a_{ij}a_{km}D_{jm}
This defines the transformation laws tensors must satisfy in rectangular geometry. The law is generalized to tensors of higher rank in the obvious way.

Now when we move to curvilinear coordinates the situation becomes more complicated. There end up being two transformation laws that are useful at various times. One is the contravariant transformation law, and the other is the covariant transformation law. Tensors can be 'mixed' in the sense that they have some components that follow one transformation and some that follow the other. The situation is complicated by the fact that the basis vectors vary from point to point in curvilinear coordinates. If you want to know about how these concepts are generalized, I will explain, but it will take a while.
 
I think you are getting confused between what a tensor is, and what the COMPONENTS of a tensor are. Things like E_{ijk} notationaly refer to the components of a tensor (0 3) tensor E, and not the tensor itself. The dot product is a linear FUNCTION on two vectors and hence IS a (0 2) tensor. Tensors are multilinear FUNCTIONS, that's it.
 
Let's start with a linear algebra review.

You know about vectors. Hopefully you're comfortable with abstract vectors spaces, but I'm just going to work with n-tuples for now.

For this entire post, I'll assume we're working up from an n-dimensional vector space.


So, a vector is simply an nx1 matrix. It has n rows and 1 column.

Then, you have covectors (a.k.a. dual vectors). In this setting, a covector is simply a 1xn matrix.

The important feature of a covector is

(covector) * (vector) = (scalar)

Then, we have nxn matrices. The important feature of nxn matrices is that

(matrix) * (vector) = (vector)

We also have some side benefits, though:

(covector) * (matrix) = (covector)

and

(covector) * (matrix) * (vector) = (scalar)



We also have another interesting feature:

(vector) * (covector) = (matrix)


This is your first nontrivial example of a tensor product.



You might wonder about other sorts of combinations such as:

(?) * (vector) = (covector)

Or even products with more than one term, like:

(?) * (vector , vector) = (scalar)

You've actually seen an example of this last thing: the dot product is a good example. Traditionally write the dot between the other two arguments:

(vector) (dot) (vector) = (scalar)

though someone fond of indices would actually write it like this:

g_{ij} v^i w^j = s

Where g is the dot, v and w are the vectors, and s is the scalar. (In this notation, you could actually write the three terms in any order you choose -- the indices specify how they're "glued together" for the product)



Anyways, in general, these more complicated things exist, and we call them tensors. A rank (p, q) tensor is something that takes p covectors and q vectors, and gives you a scalar.

For example, an nxn matrix is a rank (1,1) tensor, since we could do:

(covector) * (matrix) * (vector) = (scalar)

to produce a scalar.

A vector is a rank (1,0) tensor, since we can do:

(covector) * (vector) = (scalar)

And a covector is a rank (0,1) tensor for the same reason.

Of course, just like matrices, we can put things together in all sorts of interesting ways. A rank (1,1) tensor can operate on a rank (1,0) tensor and produce a rank (1,0) tensor, and all sorts of other stuff.


We can build higher-rank tensors out of lower-rank tensors. For example, remember the tensor product I mentioned before:

(vector) * (covector) = (matrix)

We've taken a (1,0) tensor and a (0,1) tensor and produced a (1,1) tensor!

In general, given a rank (p,q) tensor and a rank (r,s) tensor, we can take their tensor product which is a rank (p+r,q+s) tensor.


These things aren't all that fun to write out in full, but I'll give a simple example of a tensor product (\otimes is the symbol for tensor product):

<br /> \left[<br /> \begin{array}{ccc} 2 &amp; 3 &amp; 5 \end{array}<br /> \right]<br /> \otimes<br /> \left[<br /> \begin{array}{ccc} 7 &amp; 11 &amp; 13 \end{array}<br /> \right]<br /> =<br /> \left[<br /> \begin{array}{ccc|ccc|ccc}<br /> 14 &amp; 22 &amp; 26 &amp; 21 &amp; 33 &amp; 39 &amp; 35 &amp; 55 &amp; 65<br /> \end{array}<br /> \right]<br /> [/tex]<br /> <br /> That last thing is supposed to be read as a &quot;partitioned&quot; matrix -- it a 1x3 matrix, whose entries are 1x3 matrices.<br /> <br /> (actually, I may have the terms of the product backwards -- I always forget what convention people like to use when writing these things)<br /> <br /> <br /> <br /> <br /> Michael_McGovern talked a lot about transformation laws. Contrary to what he says, such things are not of fundamental importance to the notion of a tensor -- they are merely a consistency check for a particular method of using tensors.<br /> <br /> As you might recall from talk about abstract vector spaces, you cannot talk about the components of a vector until you&#039;ve selected a <u>basis</u> for that vector space. Then, you talked about comparing two different bases, and worked out a <u>change of basis</u> transformation.<br /> <br /> Most of what Michael_McGovern talked about is simply the change of basis transformations for tensors.<br /> <br /> To be fair, these things <i>are</i> important, because often times, physicists will construct the components of a tensor in some apparently basis-dependent manner -- but bases aren&#039;t supposed to matter to physics! So they have to carefully prove that their construction properly respects the change of basis transformations before they can use their tensors.<br /> <br /> <br /> <br /> Also, in physics, one is interested in <u>tensor fields</u> which is yet another layer of complexity! You&#039;ve probably heard about vector fields (such as the \vec{E} field) -- to each point of space you associate a vector. Well, you can do the same sort of thing with tensors, and have to worry about that.<br /> <br /> Since Michael_McGovern likes to do things in index notation (i.e. using coordinates), he has to select a basis. However, for technical reasons, just like a vector field assigns a vector to each point in space, you must also specify a basis for every point in space!<br /> <br /> Of course, if you are comfortable doing your linear algebra &quot;abstractly&quot; (i.e. you&#039;re happy manipulating the vector \vec{v} as opposed to insisting on picking a basis and manipulating the n-tuple (v_1, \ldots, v_n)), you avoid just about all of the issues Michael_McGovern discussed in his post.<br /> <br /> (Yes, I <i>am</i> a big fan of doing it &quot;abstractly&quot;, why do you ask? <img src="https://cdn.jsdelivr.net/joypixels/assets/8.0/png/unicode/64/1f609.png" class="smilie smilie--emoji" loading="lazy" width="64" height="64" alt=":wink:" title="Wink :wink:" data-smilie="2"data-shortname=":wink:" />)<br /> <br /> <br /> But the point I wanted to make in these closing remarks is all of these additional concerns are only concerns about how people <i>use</i> tensors (especially physicists) -- they are not concerns inherent to the tensor concept itself.
 
Last edited:
Tensor is no more than a symbol with princepal we use to express what we want express with many useful
 
That's intersesting Hurkyl. That is a much better definition of the tensor than the one I learned. I was trying to learn this stuff on my own and I got one book that was more physics-oriented and one that was more pure math. They both defined tensors in terms of the way their components transformed. I do see the advantage of your approach.
 
What are Tensors?

Hi, I have been hearing/reading the word "tensor" a lot lately, but I have no idea what it is or what is it used for. I also googled for it but I get bogged down by so much coplicated mathematics that I am unable to make any sense of it. All I know that tensors have something to do with matrices and special relativity, no more no less. Could someone please just give me a gist of what tensors are?
 
  • #10
Swapnil said:
Hi, I have been hearing/reading the word "tensor" a lot lately, but I have no idea what it is or what is it used for. I also googled for it but I get bogged down by so much coplicated mathematics that I am unable to make any sense of it. All I know that tensors have something to do with matrices and special relativity, no more no less. Could someone please just give me a gist of what tensors are?

yes, I implied this question in my vector calc thread. I didn't even bother to look it up, because I'm afraid I'll draw misconceptions from laymen explanations (as I have done in the past with relativity and quantum mechanics).
 
  • #11
i think if you will search on here over the last few years you will find thousands of words written on this question. maybe one thread was called what is a tensor?

i myself have answered this question uncountably many times.
 
  • #12
you might look in the tensor forum.
 
  • #14
Hey you guys, I am trying to learn about tensors on my own this summer and I would be really glad if someone would recommend a good book(s) on them. Preferably a book which gives you a physicist/engineer's perspective on tensors (not a mathematicaian's).

Thanks in advance.
 
  • #15
think of a taylor series expanded at each point of a space. the constant terms are the values of the function at each point. the linear etrms are the differentials of the function at each point. these are first order tensors. then the second order taylor polynomials are second order approximations to the functions ate ach point. these are second order symmetric tensors. etc...

there are also anti synmmetric tensors, like the 1-forms that one integrates over parametrized curves, and the 2 forms that one integrates over parametrized surfaces.

and there are more complicated ones. in general they are multilinear combinations of tangent vectors and cotangent vectors.
 
  • #16
mathwonk said:
think of a taylor series expanded at each point of a space.
I can't even begin to express the difficulty imagining this.

To me, Taylor series is a bunch of equations that 'zoom in' on a slope, but my only practical application with taylor series was using a runge-kutta technique to eliminate error in a computational physics class...

Other than that, it was a completely abstract equation that came at a tough time in math for me, where blatent memorization was my relief from the 'fire hose'*


*I have a physics professor who states that Physics 211/212 are like asking a student to take a drink from a fire hose. It fits with my experience. Maybe it's a local thing, but having had no physics background before 211/212, it was a crazy year; it flew through all kinds of different branches of physics while I was learning calc 2 and 3 as well. I totally lacked soak time. A lot of my mathematical concepts are severely underdeveloped.
 
  • #17
well start with one point.

i.e. think of a polynomial in two variables and collect etrms of the same total degree,like 3 + (x-4y) + (x^2 +6xy -y^2) + (x^3 +xy^2 -y^3).

the constant term, 3, is the zeroth order approximation. the lienar terms (x-4y) are the first order approximation, i.e. approximatiion by a first iorder symmetric tensor. (all a symetric tensor is, is a homogeneous polynomial).

the next etrms ((x^2 +6xy -y^2), the second order approximation, are by a second order tensor.

this polynomial is expanded in powers of X,Y.

the hard
part is when we try to expand it in powers of X-a, Y-b, for every point (a,b) in the plane. the resuklt is we get a wholelot of constants, a whole,lot of linear polynomials, a wholelot of quadratic polynomials, etc...people who do not know what tensors are, think the complicated notation that is used to express these latter families of objects are the tensors, and talk about families of coefficients with upper and lower indices as "being" tensors.

from another point of view the simplest tensors are real vaued funtions, the next simplest are vector vakued functions, and the next simplest are matrix valued functions. etcc...
 
  • #18
lets get systematic. first take the local point of view. we discuss only covariant tensors in the classical sense, and symmetric ones at that. and let us stick to 2 variables.

a covariant 0 tensor is a number. a field of these is a family of numbers, one at each point, i.e. a real valued function. ok?

now a 1 cotensor is a linear polynomial like ax+by. so a field of 1 cotensors is a family of these, one at each point. this is harder to imagine, so just imagine ax+by, but where a and b are functions,

i.e.a(p)x+b(p)y. so in a sense it is represented by the 2 coefficient functions, a(p) and b(p).

now a (symmetric) 2 cotensor is a homogeneous quadratic polynomial like ax^2 + bxy + cy^2, and thus a field of them is the same thing where a,b,c are functions. so in a sense a symmetric 2 co tensor is represented by the three coefficient functions, a,b,c.

now a 2 cotensor that may not be symmetric, will have also a yx term that cannot be combined with the xy term, so will look like ax^2 + bxy + cyx + dy^2, and a field of those is such a thing again where the coefficients are functions.

thats about it. so the big deal I gave you at first was to imagine all these degrees at once. i.e. a symmetric tensor is just a (NOT NECESSARILY HOMOGENEOUS,oops sorry) polynomiaL.


and a field of them is a polynomial same but with function coefficients. so a not necessarily symmetric cotensor is a sort of non commutative polynomial. and so on for a field of them, which then might be represented by a huge family of coefficient functions.


thus for each n, a smooth function f defines a field of symmetric tensors with top degree n, namely the field of nth order taylor polynomials at each point. the associated family of functional coefficients is just the family of partials of f up to degree n.


if we stick to a single homogeneous degree for our (still co-) tensors, we see there is one coefficient for each noncommutative monomial, i.e. in 2 variables there are the following eight degree 3 monomials xxx, xxy, xyy, yyy, yyx, yxx, xyx, yxy.


now i will probably screw this up from my allergy to coordinates, but never fear many people love those best and will leap right in with help.

but anyway, coordinate junkies will prefer to write something like (1,1,1) instead of xxx, and (1,1,2) instead of xxy, and so on, so they will describe an order three cotensor by giving only the coefficients of each term in the form {a(1,1,1), a(1,1,2),...etc...}, where since they always think of fields of cotensors, they will say the a(i,j,k) each represents a function.

hence they will say that an order 3 cotensor is a family of functions of form {a(i,j,k)} and they will also give you rules for how to change coordinates. Indeed they must do this since they have not told you what the symbols mean, so you would have no way of figuring out for yourself how they transform.

ok now i withdraw to a safe bunker to await comments from a classical tensor perspective:biggrin: .
 
Last edited:
  • #19
oh really there are three stages in understanding tensors, tensors at a point, then families of tensors in R^n, then patching together such families to get families on a manifold. i have discussedonly the first two stages.

the part many physicists leave out is stage one. they jump right into stages two and three, and hence do not understand what the objects are that they are globalizing, so are obliged to memorize the rulkes for changing coordinates instead of deriving them from conceptual definitions.

this i not their fault of course since the books they read do not explain what is going on. my near hopeless lack of grasp of physics is also due to the books i read leaving that aspect out from the math.
 
  • #20
Hey Hurkyl,

conventionally, what is the result of

\left[ \begin {array} {c} a \\ b \end {array} \right] \otimes \left[ \begin {array} {ccc} c &amp; d \end {array} \right]

??
 
  • #21
The usual definition is through what I think is called the Kronecker product. I can never remember if it's

<br /> \left[ \begin{array}{c|c}<br /> c \left[ \begin{array}{c} a \\ b \end{array} \right]<br /> &amp;<br /> d \left[ \begin{array}{c} a \\ b \end{array} \right]<br /> \end{array} \right]<br />

or

<br /> \left[<br /> \begin{array}{c}<br /> a \left[ \begin{array}{c c} c &amp; d \end{array} \right] \\<br /> \hline<br /> b \left[ \begin{array}{c c} c &amp; d \end{array} \right]<br /> \end{array}<br /> \right]<br />

Fortunately, it doesn't matter in this case -- they are both equal to

<br /> \left[<br /> \begin{array}{c c}<br /> ac &amp; ad \\<br /> bc &amp; bd<br /> \end{array}<br /> \right]<br />
 
  • #22
By the way, does anyone understand Penrose's tensor notations that look like multi-legged bugs?
 
  • #23
It's been a while since I looked at it, but I think...

The inputs to a tensor are represented as legs on the top side of the bug.
The outputs to a tensor are represented as legs on the bottom side of the bug.
Contraction is represented by connecting the wires.
Raising/lowering indices with a metric is done by attaching a U-shape to change which way (up/down) the wire is travelling.

Oh, and a tensor product is done by placing two tensors side by side.
 
Last edited:
  • #24
Simplest way I always thought you could learn what a tensor is,is by thinking of the indices/or the relationship

Vector->Matrix->Tensor
[N]->[N,M]->[N,M,L...] ...i learned it from an analy mech/programming book.

But lik hurkyl said...the vector and Matrix are also Tensors. Thus I think you can consider Tensors as a data structure accessed by indices/array operator
 
  • #25
I think that's very misleading! A matrix is not a vector nor a tensor: a matrix is just one way of representing vectors or transformations. (And most tensors cannot be written in terms of matrices.)
 
  • #26
I noted threads asking "what is a tensor" in both the General Math and Calclus sections. I am merging both threads and moving the combined thread to the "Tensor" section.
 
  • #27
thank you halls. could you also please put a little "moved" sign in the calculus section to guide the OP here?
 
  • #28
Mathwonk, could you please explain to me:

1) what is a "cotensor" as opposed to a tensor?
2) How does that view of tensors as polynomials relate to the view of tensors as functions of vectors and covectors?
 
  • #29
HallsIvy: you wrote
"For example, an nxn matrix is a rank (1,1) tensor, since we could do"
So does this imply that not all nx"m" matrices are tensors or no nxm matrices are tensors.

http://mathworld.wolfram.com/Tensor.html..mentions the whole N indices each of M-D. Does this mean that there is a unique M for all indices...or M is just a variable? so like M1...Mn
 
Last edited by a moderator:
  • #30
cotensors refer to various scalar valued functions on vectors.

they are called covariant tensors in classical language. they are the ones that pull back under mappings.


the other kind are called contravariant tensors in classical language.


lets just look at a manifold and ask what we can build from it.

1) the most fundamental object we can construct is the family of tangent spaces at each point.

a choice of a tangent vector at each point is an example of what i believe is classically called a contravariant tensor field.

2) a higher order (intellectually that is) object is the dual tangent space at each point, i.e. the spaces of linear functions on tangent vectors.

a choice of such a linear function on tangent vectors at each point is i believe an example of something classically called a covariant tensor field.


note that a linear function is nothing but (in coordinates) a linear polynomial, i.e. the linear term of a taylor series. hence assigning the linear term of the taylor series ate ach point, of a smooth function, is one way to define a covariant tensor field.


3) generalizing and raising the ante from linear to bilinear, we could consider at each point the space of bilinear real valued functions on ordered pairs of vectors.

choosing such a bilinear function at each point is another example i believe of a covariant tensor field. an example is a symmetric bilinear function at each point, i.e. in coordinates, merely a quadratic polynomial such as the second term of the taylor series of a function.

there are also non abelian quadratic polynomials, also called covariant tensors (of second order?)

4) now to be consistent we should also define a contravariant tensor field of second order, something dual to a non abelian quadratic polynomial. i skipped this because it is harder to define than the covariant version.

namely we have to define some kind of second order tangent vectors, but the definition is a little mathematical looking and less natural than the quadratic polynomials above. to be semi precise, we want some one gadget, such that evaluating a quadratic polynomial on a pair of tangent vectors, is equivalent to evaluating the 2-tensor defined by the quadratic polynomial on this one gadget.

precisely but probably unhelpfully, the tensor product of a vector space V with tiself is another vector space VtensorV, plus a bilinear map

m:VxV ----> VtensorV, such that every bilinear map VxV--W for any vector space W, can be factored as a composition

VxV--->VtensorV--->W, where the second map VtensorV--->W is linear.


it is easy to show how to write such things, just take linear combiantions of symbols like vtensorv', but that does not explain what they are.



basically just as the first order covectors, or dual vectors, are linear functions on tangent vectors, it is also nice if the second order cotensors, the polynomials, were also actually linear maps, not bilinear maps, on some vector space. VtensorV is that vector space.


i.e. second order cotensors are bilinear maps VxV--->R, but we also want to write them as (VtensorV)* = linear maps on VtensorV.

one unnatural way to do this is to simply define VtensorV = {Bil(VxV,R)}*, i.e. the dual of the bilinear maps.

since in finite dimensions, dual of a dual is the original space back, we get then that (VtensorV)* = {Bil(VxV,R)}** ={Bil(VxV,R)}.


there is another way to do it mathematically but it is so complicated i do not blame anyone for not learning this stuff abstractly.


i guess the most intuitive way to elarn is in afct the way physicists do! i.e. just elarn how to write them down, and not woprry about the definitions.

howeverm, since i am aboiut to find myself agreeing with a point of view i have foiught form years here, i will take a step back and say this:

there is a hgue difference between knowing the absrtact properties a gadget should have, and knowing the picky technical definition and construction of the agdget that ahs those proeprties.

i.e. although i do not advocate struggling throuigh the mathematical construction of tensors, i do advocate knowing the characterizing properties they have.


from a utilitarian point of view, a tensor is anything you can get by starting from a tangent space and iterating or combining the constructions of multil;inear functions on a space already given.

i.e. the dual space V* is the linear functions on V. then one can construct the bilinear maps on VxV*. this is essentially V*tensorV.

then one can take the dual of that, getting VtensorV*.

\then one can take the trilinear maps on VxVxV*, getting essentially

V*tensorV*tensorV. and so on...


cotensors are the ones with the stars on them.:bugeye:
 

Similar threads

  • · Replies 15 ·
Replies
15
Views
3K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 3 ·
Replies
3
Views
7K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 8 ·
Replies
8
Views
4K
  • · Replies 16 ·
Replies
16
Views
6K
  • · Replies 2 ·
Replies
2
Views
4K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 6 ·
Replies
6
Views
3K