Sierpinski triangle: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>MusikAnimal
m Reverted good faith edits by 71.178.27.45 talk: Discospinster (HG)
Undid revision 595232546 by AndrewProgrammer (talk) - original research
Line 1: Line 1:
{{other uses}}
Hello and welcome. My name is Figures Wunder. The favorite pastime for my children and me is to perform baseball but I haven't made a dime with it. Bookkeeping is her working day job now. Years ago we moved to North Dakota.<br><br>My web page: [http://Www.Garuda.org//weightlossfooddelivery24650 healthy meals delivered]
 
[[File:Components stress tensor.svg|right|thumb|300px|[[Cauchy stress tensor]], a second-order tensor. The tensor's components, in a three-dimensional Cartesian coordinate system, form the matrix<br />
<math>\begin{align}
\sigma & = \begin{bmatrix}\mathbf{T}^{(\mathbf{e}_1)} \mathbf{T}^{(\mathbf{e}_2)} \mathbf{T}^{(\mathbf{e}_3)} \\ \end{bmatrix} \\
& = \begin{bmatrix} \sigma_{11} & \sigma_{12} & \sigma_{13} \\ \sigma_{21} & \sigma_{22} & \sigma_{23} \\ \sigma_{31} & \sigma_{32} & \sigma_{33} \end{bmatrix}\\
\end{align}</math><br />
whose columns are the stresses (forces per unit area) acting on the '''e'''<sub>1</sub>, '''e'''<sub>2</sub>, and '''e'''<sub>3</sub> faces of the cube.]]
 
'''Tensors''' are [[geometry|geometric]] objects that describe [[linear relation]]s between [[Euclidean vectors|vectors]], [[Scalar (mathematics)|scalars]], and other tensors. Elementary examples of such relations include the [[dot product]], the [[cross product]], and [[linear map]]s. Vectors and scalars themselves are also tensors. A tensor can be represented as a [[Array data structure#Multidimensional arrays|multi-dimensional array]] of numerical values. The '''{{visible anchor|order}}''' (also ''degree'') of a tensor is the dimensionality of the array needed to represent it, or equivalently, the number of indices needed to label a component of that array. For example, a linear map can be represented by a matrix, a 2-dimensional array, and therefore is a 2nd-order tensor. A vector can be represented as a 1-dimensional array and is a 1st-order tensor. Scalars are single numbers and are thus 0th-order tensors.
 
Tensors are used to represent correspondences between sets of [[Euclidean vector|geometric vectors]].  For example, the [[Cauchy stress tensor]] '''T''' takes a direction '''v''' as input and produces the stress '''T'''<sup>(''v'')</sup> on the surface normal to this vector for output thus expressing a relationship between these two vectors, shown in the figure (right).
 
Because they express a relationship between vectors, tensors themselves must be [[coordinate-free|independent of a particular choice]] of [[coordinate system]]. Taking a coordinate [[basis of a vector space|basis]] or [[frame of reference]] and applying the tensor to it results in an organized multidimensional array representing the tensor in that basis, or frame of reference.  The coordinate independence of a tensor then takes the form of a [[covariant transformation|"covariant" transformation law]] that relates the array computed in one coordinate system to that computed in another one.  This transformation law is considered to be built into the notion of a tensor in a geometric or physical setting, and the precise form of the transformation law determines the ''type'' (or ''valence'') of the tensor.
 
Tensors are important in physics because they provide a concise mathematical framework for formulating and solving physics problems in areas such as elasticity, fluid mechanics, and general relativity.  Tensors were first conceived by [[Tullio Levi-Civita]] and [[Gregorio Ricci-Curbastro]], who continued the earlier work of [[Bernhard Riemann]] and [[Elwin Bruno Christoffel]] and others, as part of the ''absolute differential calculus''. The concept enabled an alternative formulation of the intrinsic [[differential geometry]] of a [[manifold]] in the form of the [[Riemann curvature tensor]].<ref name="Kline">{{cite book|title=Mathematical thought from ancient to modern times, Vol. 3|first=Morris|last=Kline|pages=1122&ndash;1127|publisher=Oxford University Press|year=1972|isbn= 0195061373}}</ref>
 
==History==
The concepts of later tensor analysis arose from the work of [[Carl Friedrich Gauss]] in differential geometry, and the formulation was much influenced by the theory of [[algebraic form]]s and invariants developed during the middle of the nineteenth century.<ref>{{cite book
|first=Karin |last=Reich
|title=Die Entwicklung des Tensorkalküls
|year=1994
|publisher=Birkhäuser
|isbn=978-3-7643-2814-6
|series=Science networks historical studies, v. 11
 
|oclc= 31468174
|url=http://books.google.com/books?id=O6lixBzbc0gC
}}</ref> The word "tensor" itself was introduced in 1846 by [[William Rowan Hamilton]]<ref>{{cite journal
|first=William Rowan |last=Hamilton
|title=On some Extensions of Quaternions
|url=http://www.emis.de/classics/Hamilton/ExtQuat.pdf
|journal=Philosophical Magazine
|year=1854–1855
|pages=492–499, 125–137, 261–269, 46–51, 280–290
|editor-first=David R.|editor-last=Wilkins
|issue=7–9
 
|issn=0302-7597
}}</ref> to describe something different from what is now meant by a tensor.<ref group=Note>Namely, the [[norm (mathematics)|norm operation]] in a certain type of algebraic system (now known as a [[Clifford algebra]]).</ref> The contemporary usage was brought in by [[Woldemar Voigt]] in 1898.<ref>{{cite book
|first=Woldemar |last=Voigt
|title=Die fundamentalen physikalischen Eigenschaften der Krystalle in elementarer Darstellung
|publisher=Von Veit
 
|place=Leipzig
|year=1898}}
</ref>
 
Tensor calculus was developed around 1890 by [[Gregorio Ricci-Curbastro]] under the title ''absolute differential calculus'', and originally presented by Ricci in 1892.<ref>{{cite journal
|first=G. |last=Ricci Curbastro
|title=Résumé de quelques travaux sur les systèmes variables de fonctions associés à une forme différentielle quadratique
|journal=Bulletin des Sciences Mathématiques
|volume=2
|pages=167–189
|year=1892
|issue=16
}}</ref> It was made accessible to many mathematicians by the publication of Ricci and [[Tullio Levi-Civita]]'s 1900 classic text ''Méthodes de calcul différentiel absolu et leurs applications'' (Methods of absolute differential calculus and their applications).<ref>{{Harv|Ricci|Levi-Civita|1900}}</ref>
 
In the 20th century, the subject came to be known as ''tensor analysis'', and achieved broader acceptance with the introduction of [[Albert Einstein|Einstein]]'s theory of [[general relativity]], around 1915. General relativity is formulated completely in the language of tensors. Einstein had learned about them, with great difficulty, from the geometer [[Marcel Grossmann]].<ref>{{cite book
|first=Abraham |last=Pais
|title=Subtle Is the Lord: The Science and the Life of Albert Einstein
|publisher=Oxford University Press
|year=2005
|isbn=978-0-19-280672-7
|url=http://books.google.com/books/about/Subtle_is_the_Lord.html?id=U2mO4nUunuwC
}}</ref> Levi-Civita then initiated a correspondence with Einstein to correct mistakes Einstein had made in his use of tensor analysis. The correspondence lasted 1915–17, and was characterized by mutual respect:
 
{{quote|I admire the elegance of your method of computation; it must be nice to ride through these fields upon the horse of true mathematics while the like of us have to make our way laboriously on foot.|Albert Einstein|The Italian Mathematicians of Relativity<ref name="Goodstein">{{cite journal
|last=Goodstein|first=Judith R
|title = The Italian Mathematicians of Relativity
|journal = Centaurus
|volume = 26
|doi = 10.1111/j.1600-0498.1982.tb00665.x
|pages = 241–261
|year = 1982
|bibcode = 1982Cent...26..241G
|issue = 3 }}</ref>}}
 
Tensors were also found to be useful in other fields such as [[continuum mechanics]]. Some well-known examples of tensors in [[differential geometry]] are [[quadratic form]]s such as [[metric tensor]]s, and the [[Riemann curvature tensor]]. The [[exterior algebra]] of [[Hermann Grassmann]], from the middle of the nineteenth century, is itself a tensor theory, and highly geometric, but it was some time before it was seen, with the theory of [[differential form]]s, as naturally unified with tensor calculus. The work of [[Élie Cartan]] made differential forms one of the basic kinds of tensors used in mathematics.
 
From about the 1920s onwards, it was realised that tensors play a basic role in [[algebraic topology]] (for example in the [[Künneth theorem]]).{{Citation needed|date=September 2011}} Correspondingly there are types of tensors at work in many branches of [[abstract algebra]], particularly in [[homological algebra]] and [[representation theory]]. Multilinear algebra can be developed in greater generality than for scalars coming from a [[field (mathematics)|field]], but the theory is then certainly less geometric, and computations more technical and less algorithmic.{{Clarify|date=April 2010}} Tensors are generalized within [[category theory]] by means of the concept of [[monoidal category]], from the 1960s.
 
==Definition==
There are several approaches to defining tensors. Although seemingly different, the approaches just describe the same geometric concept using different languages and at different levels of abstraction.
 
===As multidimensional arrays===
 
Just as a scalar is described by a single number, and a vector with respect to a given basis is described [[Euclidean vector#Representations|by an array]] of one dimension, any tensor with respect to a basis is described by a multidimensional array. The numbers in the array are known as the ''scalar components'' of the tensor or simply its ''components.'' They are denoted by indices giving their position in the array, in [[subscript and superscript]], after the symbolic name of the tensor. The total number of indices required to uniquely select each component is equal to the dimension of the array, and is called the ''order'' or the ''rank'' of the tensor.<ref group="Note">This article will be using the term ''order'', since the term ''rank'' has a different meaning in the related context of matrices.</ref> For example, the entries of an order 2 tensor ''T'' would be denoted ''T''<sub>''ij''</sub>, where ''i'' and ''j'' are indices running from 1 to the dimension of the related vector space.<ref group="Note">Vector spaces in this article are assumed to be finite dimensional, unless otherwise noted.</ref>
 
Just as the components of a vector change when we change the [[basis (linear algebra)|basis]] of the vector space, the entries of a tensor also change under such a transformation. Each tensor comes equipped with a ''transformation law'' that details how the components of the tensor respond to a [[change of basis]]. The components of a vector can respond in two distinct ways to a [[change of basis]] (see [[covariance and contravariance of vectors]]), where the new basis vectors <math>\mathbf{\hat{e}}_i </math> are expressed in terms of the old basis vectors  <math>\mathbf{e}_j </math> as,
:<math>\mathbf{\hat{e}}_i = \sum_j R^j_i \mathbf{e}_j = R^j_i \mathbf{e}_j,</math>
where ''R''<sub>''i''</sub><sup>'' j''</sup> is a matrix and in the second expression the summation sign was suppressed (a notational convenience [[Einstein summation convention|introduced by Einstein]] that will be used throughout this article). The components, ''v''<sup>''i''</sup>, of a regular (or column) vector, '''v''', transform with the [[matrix inverse|inverse]] of the matrix ''R'',
:<math>\hat{v}^i = (R^{-1})^i_j v^j,</math>
where the hat denotes the components in the new basis. While the components, ''w''<sub>''i''</sub>, of a covector (or row vector), '''w''' transform with the matrix R itself,
:<math>\hat{w}_i = R_i^j w_j.</math>
The components of a tensor transform in a similar manner with a transformation matrix for each index. If an index transforms like a vector with the inverse of the basis transformation, it is called ''contravariant'' and is traditionally denoted with an upper index, while an index that transforms with the basis transformation itself is called ''covariant'' and is denoted with a lower index. The transformation law for an order-''m'' tensor with ''n'' contravariant indices and ''m''−''n'' covariant indices is thus given as,
:<math>\hat{T}^{i_1,\ldots,i_n}_{i_{n+1},\ldots,i_m}= (R^{-1})^{i_1}_{j_1}\cdots(R^{-1})^{i_n}_{j_n} R^{j_{n+1}}_{i_{n+1}}\cdots R^{j_{m}}_{i_{m}}T^{j_1,\ldots,j_n}_{j_{n+1},\ldots,j_m}.</math>
Such a tensor is said to be of order or ''type'' {{Nowrap|(''n'',''m''−''n'')}}.<ref group="Note">There is a plethora of different terminology for this around. The terms "order", "type", "rank", "valence", and "degree" are in use for the same concept. This article uses the term "order" or "total order" for the total dimension of the array (or its generalisation in other definitions) ''m'' in the preceding example, and the term "type" for the pair giving the number contravariant and covariant indices. A tensor of type {{Nowrap|(''n'',''m''−''n'')}} will also be referred to as a "{{Nowrap|(''n'',''m''−''n'')}}" tensor for short.</ref>
This discussion motivates the following formal definition:<ref>{{cite book | last1=Sharpe | first1=R. W. | title=Differential Geometry: Cartan's Generalization of Klein's Erlangen Program | publisher=[[Springer-Verlag]] | location=Berlin, New York | isbn=978-0-387-94732-7 | year=1997|page=194}}</ref>
 
{{quotation|'''Definition.'''  A tensor of type (''n'', ''m''−''n'') is an assignment of a multidimensional array
:<math>T^{i_1\dots i_n}_{i_{n+1}\dots i_m}[\mathbf{f}]</math>
to each basis {{math|'''f''' {{=}} ('''e'''<sub>1</sub>,...,'''e'''<sub>''N''</sub>)}} such that, if we apply the change of basis
:<math>\mathbf{f}\mapsto \mathbf{f}\cdot R = \left( R_1^i \mathbf{e}_i, \dots, R_N^i\mathbf{e}_i\right)</math>
then the multidimensional array obeys the transformation law
:<math>T^{i_1\dots i_n}_{i_{n+1}\dots i_m}[\mathbf{f}\cdot R] = (R^{-1})^{i_1}_{j_1}\cdots(R^{-1})^{i_n}_{j_n} R^{j_{n+1}}_{i_{n+1}}\cdots R^{j_{m}}_{i_{m}}T^{j_1,\ldots,j_n}_{j_{n+1},\ldots,j_m}[\mathbf{f}].</math>
}}
 
The definition of a tensor as a multidimensional array satisfying a transformation law traces back to the work of Ricci.<ref name="Kline"/> Nowadays, this definition is still used in some physics and engineering text books.<ref>{{cite book
|title=Classical Dynamics of Particles and Systems
|last1=Marion
|first1=J.B.
|last2=Thornton
|first2=S.T.
|page=424
|edition=4th
|year=1995
|publisher=Saunders College Publishing
|isbn=978-0-03-098967-4
}}</ref><ref>{{cite book
|title=Introduction to Electrodynamics
|last1=Griffiths
|first1=D.J.
|isbn=978-0-13-805326-0
|edition=3
|pages=11–12 and 535–
|publisher=Prentice Hall
|year=1999
}}</ref>
 
====Tensor fields====
{{Main|Tensor field}}
In many applications, especially in differential geometry and physics, it is natural to consider a tensor with components which are functions. This was, in fact, the setting of Ricci's original work. In modern mathematical terminology such an object is called a [[tensor field]], but they are often simply referred to as tensors themselves.<ref name="Kline"/>
 
In this context the defining transformation law takes a different form. The "basis" for the tensor field is determined by the coordinates of the underlying space, and the defining transformation law is expressed in terms of [[partial derivative]]s of the coordinate functions, <math>\bar{x}_i(x_1,\ldots,x_k)</math>, defining a coordinate transformation,<ref name="Kline"/>
:<math>\hat{T}^{i_1\dots i_n}_{i_{n+1}\dots i_m}(\bar{x}_1,\ldots,\bar{x}_k) =
\frac{\partial \bar{x}^{i_1}}{\partial x^{j_1}}
\cdots
\frac{\partial \bar{x}^{i_n}}{\partial x^{j_n}}
\frac{\partial x^{j_{n+1}}}{\partial \bar{x}^{i_{n+1}}}
\cdots
\frac{\partial x^{j_m}}{\partial \bar{x}^{i_m}}
T^{j_1\dots j_n}_{j_{n+1}\dots j_m}(x_1,\ldots,x_k).
</math>
 
===As multilinear maps===
A downside to the definition of a tensor using the multidimensional array approach is that it is not apparent from the definition that the defined object is indeed basis independent, as is expected from an intrinsically geometric object. Although it is possible to show that transformation laws indeed ensure independence from the basis, sometimes a more intrinsic definition is preferred. One approach <!--introduced by ???, -->is to define a tensor as a [[multilinear map]]. In that approach a type (''n'',''m'') tensor ''T'' is defined as a map,
:<math> T: \underbrace{ V^* \times\dots\times V^*}_{n \text{ copies}} \times \underbrace{ V \times\dots\times V}_{m \text{ copies}} \rightarrow \mathbf{R}, </math>
where ''V'' is a [[vector space]] and ''V''* is the corresponding [[dual space]] of covectors, which is linear in each of its arguments.
 
By applying a multilinear map ''T'' of type (''n'',''m'') to a basis {'''e'''<sub>j</sub>} for ''V'' and a canonical cobasis {'''ε'''<sup>i</sup>} for ''V''*,
:<math>T^{i_1\dots i_n}_{j_1\dots j_m} \equiv T(\mathbf{\varepsilon}^{i_1},\ldots,\mathbf{\varepsilon}^{i_n},\mathbf{e}_{j_1},\ldots,\mathbf{e}_{j_m}),</math>
an ''n''+''m'' dimensional array of components can be obtained. A different choice of basis will yield different components. But, because ''T'' is linear in all of its arguments, the components satisfy the tensor transformation law used in the multilinear array definition. The multidimensional array of components of ''T'' thus form a tensor according to that definition. Moreover, such an array can be realised as the components of some multilinear map ''T''. This motivates viewing multilinear maps as the intrinsic objects underlying tensors.
 
===Using tensor products===
{{Main|Tensor (intrinsic definition)}}
For some mathematical applications, a more abstract approach is sometimes useful. This can be achieved by defining tensors in terms of elements of [[tensor product]]s of vector spaces, which in turn are defined through a [[universal property]]. A type (''n'',''m'') tensor is defined in this context as an element of the tensor product of vector spaces,<ref>{{Springer|id=a/a011120|title=Affine tensor}}</ref>
:<math> T\in \underbrace{V \otimes\dots\otimes V}_{n \text{ copies}} \otimes \underbrace{V^* \otimes\dots\otimes V^*}_{m \text{ copies}}.</math>
 
If '''v'''<sub>i</sub> is a basis of ''V'' and '''w'''<sub>j</sub> is a basis of ''W'', then the tensor product <math> V\otimes W</math> has a natural basis <math> \mathbf{v}_i\otimes \mathbf{w}_j</math>. The components of a tensor ''T'' are the coefficients of the tensor with respect to the basis obtained from a basis {'''e'''<sub>i</sub>} for ''V'' and its dual {'''ε'''<sup>j</sup>}, i.e.
:<math>T = T^{i_1\dots i_n}_{j_1\dots j_m}\; \mathbf{e}_{i_1}\otimes\cdots\otimes \mathbf{e}_{i_n}\otimes \mathbf{\varepsilon}^{j_1}\otimes\cdots\otimes \mathbf{\varepsilon}^{j_m}.</math>
Using the properties of the tensor product, it can be shown that these components satisfy the transformation law for a type (''m'',''n'') tensor. Moreover, the universal property of the tensor product gives a [[bijection|1-to-1 correspondence]] between tensors defined in this way and tensors defined as multilinear maps.
 
==Examples==
{{see also|Dyadic tensor}}
<!--NOTE: "Dyadic" is old terminology for matrix representations of rank-2 tensors which could be any of (2,0), (1,1) or (0,2) (which clashes with bilinear forms, inner products, metric etc.) - its meaning is vague, unclear and confusing, and they aren't really used anymore. Please do not add "dyadic tensor" to the table. Thanks. -->
{{multiple image
  | right
  | footer    = Geometric interpretations for [[exterior product]]s of ''n'' [[vector (geometry)|vector]]s and ''n'' [[1-form]]s, where ''n'' is the [[graded algebra|grade]],<ref>{{cite book |author=R. Penrose| title=[[The Road to Reality]]| publisher= Vintage books| year=2007 | isbn=0-679-77631-1}}</ref> for ''n'' = 1, 2, 3. The "circulations" show [[Orientation (vector space)|orientation]].<ref>{{cite book|title=Gravitation|author=J.A. Wheeler, C. Misner, K.S. Thorne|publisher=W.H. Freeman & Co|year=1973|page=83|isbn=0-7167-0344-0}}</ref>
  | width1    = 125
  | image1    = N-vector.svg
  | caption1  = [[vector (geometry)|Vector]]s ('''u''', '''v''', '''w''') exterior-multiplied to obtain [[Multivector|''n''-vector]]s ([[parallelotope]] elements).
  | width2    = 125
  | image2    = N-form.svg
  | caption2  = 1-forms ('''ε''', '''η''', '''ω''') exterior-multiplied to obtain ''n''-forms ("meshes" of [[coordinate surface]]s, here planes)
  }}
 
This table shows important examples of tensors, including both tensors on vector spaces and tensor fields on manifolds. The tensors are classified according to their type (''n'', ''m''). For example, a [[bilinear form]] is the same thing as a (0, 2)-tensor; an [[inner product]] is an example of a (0, 2)-tensor, but not all (0, 2)-tensors are inner products. In the (0, ''M'')-entry of the table, ''M'' denotes the dimension of the underlying vector space or manifold.
 
:{| class = "wikitable"
|-
!scope="col" = width="75px" | ''n, m''
!scope="col" = width="175px" | ''n'' = 0
!scope="col" = width="175px" | ''n'' = 1
!scope="col" = width="175px" | ''n'' = 2
!scope="col" = width="75px" | ...
!scope="col" = width="175px" | ''n'' = ''N''
!scope="col" = width="75px" | ...
|-
!''m'' = 0
|[[scalar (mathematics)|scalar]], e.g. [[scalar curvature]]
|[[Euclidean vector|vector]] (e.g. [[direction vector]])
|[[bivector]], e.g. inverse metric tensor
|
| [[Multivector|''N''-vector]], a sum of [[Blade (geometry)|''N''-blade]]s
|
|-
!''m'' = 1
|[[covector]], [[linear functional]], [[1-form]] (e.g. [[gradient]] of a scalar field)
|[[linear transformation]], [[Kronecker delta]]
|
|
|
|
|-
!''m'' = 2
|[[bilinear form]], e.g. [[inner product]], [[metric tensor]], [[Ricci curvature]], [[2-form]], [[symplectic form]]
|e.g. [[cross product]] in three dimensions
|e.g. [[elasticity tensor]]
|
|
|
|-
!''m'' = 3
|e.g. 3-form
|e.g. [[Riemann curvature tensor]]
|
|
|
|
|-
!...
|
|
|
|
|
|
|-
!''m'' = ''M''
|e.g. ''M''-form i.e. [[volume form]]
|
|
|
|
|
|-
!...
|
|
|
|
|
|
|-
|}
 
Raising an index on an (''n'', ''m'')-tensor produces an (''n'' + 1, ''m'' − 1)-tensor; this can be visualized as moving diagonally up and to the right on the table. Symmetrically, lowering an index can be visualized as moving diagonally down and to the left on the table. [[#Contraction|Contraction]] of an upper with a lower index of an (''n'', ''m'')-tensor produces an (''n'' − 1, ''m'' − 1)-tensor; this can be visualized as moving diagonally up and to the left on the table.
 
==Notation==
 
===Ricci calculus===
 
[[Ricci calculus]] is the modern formalism and notation for tensor indices: indicating [[inner product|inner]] and [[outer product]]s, [[covariance and contravariance of vectors|covariance and contravariance]], [[summation]]s of tensor components, [[symmetric tensor|symmetry]] and [[antisymmetric tensor|antisymmetry]], and [[partial derivative|partial]] and [[covariant derivative]]s.
 
===Einstein summation convention===
 
The [[Einstein summation convention]] dispenses with writing [[summation sign]]s, leaving the summation implicit. Any repeated index symbol is summed over: if the index ''i'' is used twice in a given term of a tensor expression, it means that the term is to be summed for all ''i''. Several distinct pairs of indices may be summed this way.
 
===Penrose graphical notation===
 
[[Penrose graphical notation]] is a diagrammatic notation which replaces the symbols for tensors with shapes, and their indices by lines and curves. It is independent of basis elements, and requires no symbols for the indices.
 
===Abstract index notation===
 
The [[abstract index notation]] is a way to write tensors such that the indices are no longer thought of as numerical, but rather are [[Indeterminate (variable)|indeterminate]]s.  This notation captures the expressiveness of indices and the basis-independence of index-free notation.
 
===Component-free notation===
 
A [[component-free treatment of tensors]] uses notation that emphasises that tensors do not rely on any basis, and is defined in terms of the [[Tensor product#Tensor product of vector_spaces|tensor product of vector spaces]].
 
==Operations==
There are a number of basic operations that may be conducted on tensors that again produce a tensor. The linear nature of tensor implies that two tensors of the same type may be added together, and that tensors may be multiplied by a scalar with results analogous to the [[Scalar multiplication|scaling of a vector]]. On components, these operations are simply performed component for component. These operations do not change the type of the tensor, however there also exist operations that change the type of the tensors.
 
===Tensor product===
{{Main|Tensor product}}
 
The [[tensor product]] takes two tensors, ''S'' and ''T'', and produces a new tensor, ''S'' ⊗ ''T'', whose order is the sum of the orders of the original tensors. When described as multilinear maps, the tensor product simply multiplies the two tensors, i.e.
:<math>(S\otimes T)(v_1,\ldots, v_n, v_{n+1},\ldots, v_{n+m}) = S(v_1,\ldots, v_n)T( v_{n+1},\ldots, v_{n+m}),</math>
which again produces a map that is linear in all its arguments. On components the effect similarly is to multiply the components of the two input tensors, i.e.
 
:<math>(S\otimes T)^{i_1\ldots i_l i_{l+1}\ldots i_{l+n}}_{j_1\ldots j_k j_{k+1}\ldots j_{k+m}} =
S^{i_1\ldots i_l}_{j_1\ldots j_k} T^{i_{l+1}\ldots i_{l+n}}_{j_{k+1}\ldots j_{k+m}},</math>
If ''S'' is of type (''l'',''k'') and ''T'' is of type (''n'',''m''), then the tensor product  ''S'' ⊗ ''T'' has type (''l''+''n'',''k''+''m'').
 
===Contraction===
{{Main|Tensor contraction}}
 
[[Tensor contraction]] is an operation that reduces the total order of a tensor by two. More precisely, it reduces a type (''n'',''m'') tensor to a type (''n''−1,''m''−1) tensor. In terms of components, the operation is achieved by summing over one contravariant and one covariant index of tensor. For example, a (1,1)-tensor <math>T_i^j</math> can be contracted to a scalar through
:<math>T_i^i</math>.
Where the summation is again implied. When the (1,1)-tensor is interpreted as a linear map, this operation is known as the [[trace (linear algebra)|trace]].
 
The contraction is often used in conjunction with the tensor product to contract an index from each tensor.
 
The contraction can also be understood in terms of the definition of a tensor as an element of a tensor product of copies of the space ''V'' with the space ''V''<sup>*</sup> by first decomposing the tensor into a linear combination of simple tensors, and then applying a factor from ''V''<sup>*</sup> to a factor from ''V''.  For example, a tensor
:<math>T \in V\otimes V\otimes V^*</math>
can be written as a linear combination
:<math>T=v_1\otimes w_1\otimes \alpha_1 + v_2\otimes w_2\otimes \alpha_2 +\cdots + v_N\otimes w_N\otimes \alpha_N.</math>
The contraction of ''T'' on the first and last slots is then the vector
:<math>\alpha_1(v_1)w_1 + \alpha_2(v_2)w_2+\cdots+\alpha_N(v_N)w_N.</math>
 
===Raising or lowering an index===
{{Main|Raising and lowering indices}}
When a vector space is equipped with an [[inner product]] (or ''metric'' as it is often called in this context), operations can be defined that convert a contravariant (upper) index into a covariant (lower) index and vice versa. A metric itself is a (symmetric) (0,2)-tensor, it is thus possible to contract an upper index of a tensor with one of lower indices of the metric. This produces a new tensor with the same index structure as the previous, but with lower index in the position of the contracted upper index. This operation is quite graphically known as ''lowering an index''.
 
Conversely the matrix inverse of the metric can be defined, which behaves as a (2,0)-tensor. This ''inverse metric'' can be contracted with a lower index to produce an upper index. This operation is called ''raising an index''.
 
==Applications==
 
===Continuum mechanics===
 
Important examples are provided by continuum mechanics. The stresses inside a [[solid body]] or [[fluid]] are described by a tensor. The [[Stress (mechanics)|stress tensor]] and [[strain tensor]] are both second-order tensors, and are related in a general linear elastic material by a fourth-order [[elasticity tensor]]. In detail, the tensor quantifying stress in a 3-dimensional solid object has components that can be conveniently represented as a 3×3 array. The three faces of a cube-shaped infinitesimal volume segment of the solid are each subject to some given force. The force's vector components are also three in number. Thus, 3×3, or 9 components are required to describe the stress at this cube-shaped infinitesimal segment. Within the bounds of this solid is a whole mass of varying stress quantities, each requiring 9 quantities to describe. Thus, a second-order tensor is needed.
 
If a particular [[Volume form|surface element]] inside the material is singled out, the material on one side of the surface will apply a force on the other side. In general, this force will not be orthogonal to the surface, but it will depend on the orientation of the surface in a linear manner. This is described by a tensor of [[type of a tensor|type (2,0)]], in [[linear elasticity]], or more precisely by a tensor field of type (2,0), since the stresses may vary from point to point.
 
===Other examples from physics===
Common applications include
 
* [[Electromagnetic tensor]] (or Faraday's tensor) in [[electromagnetism]]
* [[Finite deformation tensors]] for describing deformations and [[strain tensor]] for [[Strain (materials science)|strain]] in [[continuum mechanics]]
* [[Permittivity]] and [[electric susceptibility]] are tensors in [[anisotropic]] media
* [[Four-tensors]] in [[general relativity]] (e.g. [[stress-energy tensor]]), used to represent [[momentum]] [[flux]]es
* Spherical tensor operators are the eigenfunctions of the quantum [[angular momentum operator]] in [[spherical coordinates]]
* Diffusion tensors, the basis of [[Diffusion Tensor Imaging]], represent rates of diffusion in biologic environments
* [[Quantum Mechanics]] and [[Quantum Computing]] utilise tensor products for combination of quantum states
 
===Applications of tensors of order > 2===
The concept of a tensor of order two is often conflated with that of a matrix. Tensors of higher order do however capture ideas important in science and engineering, as has been shown successively in numerous areas as they develop. This happens, for instance, in the field of [[computer vision]], with the [[trifocal tensor]] generalizing the [[fundamental matrix (computer vision)|fundamental matrix]].
 
The field of [[nonlinear optics]] studies the changes to material [[Polarization_density#Relation between P and E in various materials|polarization density]] under extreme electric fields. The polarization waves generated are related to the generating [[electric field]]s through the nonlinear susceptibility tensor. If the polarization '''P''' is not linearly proportional to the electric field '''E''', the medium is termed ''nonlinear''.  To a good approximation (for sufficiently weak fields, assuming no permanent dipole moments are present), '''P''' is given by a [[Taylor series]] in '''E''' whose coefficients are the nonlinear susceptibilities:
 
:<math> \frac{P_i}{\varepsilon_0} = \sum_j  \chi^{(1)}_{ij} E_j  +  \sum_{jk} \chi_{ijk}^{(2)} E_j E_k + \sum_{jk\ell} \chi_{ijk\ell}^{(3)} E_j E_k E_\ell  + \cdots. \!</math>
 
Here <math>\chi^{(1)}</math> is the linear susceptibility, <math>\chi^{(2)}</math> gives the [[Pockels effect]] and  [[second harmonic generation]], and <math>\chi^{(3)}</math> gives the [[Kerr effect]]. This expansion shows the way higher-order tensors arise naturally in the subject matter.
 
==Generalizations==
 
===Tensors in infinite dimensions===
The notion of a tensor can be generalized in a variety of ways to [[Dimension (vector space)|infinite dimensions]].  One, for instance, is via the [[tensor product of Hilbert spaces|tensor product]] of [[Hilbert space]]s.<ref>{{cite jstor|1992855}}</ref> Another way of generalizing the idea of tensor, common in [[Nonlinear system|nonlinear analysis]], is via the [[#As multilinear maps|multilinear maps definition]] where instead of using finite-dimensional vector spaces and their [[algebraic dual]]s, one uses infinite-dimensional [[Banach space]]s and their [[continuous dual]].<ref>{{cite book
|last1=Abraham
|first1=Ralph
|authorlink1=
|last2=Marsden
|first2=Jerrold E.
|last3=Ratiu
|first3=Tudor S.
|editor1-first=
|editor1-last=
|editor1-link=
|others=
|title=Manifolds, Tensor Analysis and Applications
|trans_title=
|url=
|format=
|accessdate=
|type=
|edition=2nd
|series=Applied Mathematical Sciences, v. 75
|volume=75
|date=
|year=1988
|month=February
|origyear=First Edition 1983
|publisher=Springer-Verlag
|location=New York
|isbn=0-387-96790-7
|oclc= 18562688
|doi= |id=
|page=
|pages=338–339
|at=
|trans_chapter=
|chapter=Chapter 5 Tensors
|chapterurl=
|quote=Elements of T<sup>r</sup><sub>s</sub> are called tensors on E,
|ref=
|bibcode=
|laysummary=
|laydate=
|separator=
|postscript=
|lastauthoramp=}}</ref>  Tensors thus live naturally on [[Banach manifold]]s.<ref>{{cite book | last1=Lang | first1=Serge | author1-link=Serge Lang | title=Differential manifolds | publisher=[[Addison-Wesley]] Pub. Co. | year=1972 |isbn= 0201041669 |location=Reading, Mass.}}</ref>
 
===Tensor densities===
{{Main|Tensor density}}
It is also possible for a [[tensor field]] to have a "density". A tensor with density ''r'' transforms as an ordinary tensor under coordinate transformations, except that it is also multiplied by the determinant of the [[Jacobian matrix and determinant|Jacobian]] to the ''r''<sup>th</sup> power.<ref>{{Springer|id=T/t092390|title=Tensor density}}</ref> Invariantly, in the language of multilinear algebra, one can think of tensor densities as [[multilinear map]]s taking their values in a [[density bundle]] such as the (1-dimensional) space of ''n''-forms (where ''n'' is the dimension of the space), as opposed to taking their values in just '''R'''. Higher "weights" then just correspond to taking additional tensor products with this space in the range.
 
In the language of [[vector bundle]]s, the determinant bundle of the [[tangent bundle]] is a [[line bundle]] that can be used to 'twist' other bundles ''r'' times. While locally the more general transformation law can indeed be used to recognise these tensors, there is a global question that arises, reflecting that in the transformation law one may write either the Jacobian determinant, or its absolute value. Non-integral powers of the (positive) transition functions of the bundle of densities make sense, so that the weight of a density, in that sense, is not restricted to integer values.
 
Restricting to changes of coordinates with positive Jacobian determinant is possible on [[orientable manifold]]s, because there is a consistent global way to eliminate the minus signs; but otherwise the line bundle of densities and the line bundle of ''n''-forms are distinct. For more on the intrinsic meaning, see [[density on a manifold]].)
 
===Spinors===
{{Main|Spinor}}
Starting with an [[Orthonormality|orthonormal]] coordinate system, a tensor transforms in a certain way when a rotation is applied.  However, there is additional structure to the group of rotations that is not exhibited by the transformation law for tensors: see [[orientation entanglement]] and [[plate trick]].  Mathematically, the [[special orthogonal group|rotation group]] is not [[simply connected]].  [[Spinor]]s are mathematical objects that generalize the transformation law for tensors in a way that is sensitive to this fact.
 
==See also==
 
===Foundational===
* [[Cartesian tensor]]
* [[Fibre bundle]]
* [[Glossary of tensor theory]]
* [[Multilinear_subspace_learning#Multilinear_projection|Multilinear projection]]
* [[One-form]]
* [[Tensor product of modules]]
 
===Applications===
* [[Application of tensor theory in engineering]]
* [[Covariant derivative]]
* [[Curvature]]
* [[Diffusion MRI#Mathematical foundation—tensors|Diffusion tensor MRI]]
* [[Einstein field equations]]
* [[Fluid mechanics]]
* [[Multilinear subspace learning]]
* [[Riemannian geometry]]
* [[Structure Tensor]]
* [[Tensor decomposition]]
* [[Tensor derivative]]
* [[Tensor software]]
 
==Notes==
{{Reflist|group="Note"}}
 
==References==
;General
{{Refbegin|30em}}
*{{cite book
| last = Bishop
| first = Richard L.
| coauthors = Samuel I. Goldberg
| title = Tensor Analysis on Manifolds
| year=  1980
| publisher = Dover
| isbn = 978-0-486-64039-6
| origyear = 1968
}}
*{{cite book
| last = Danielson
| first = Donald A.
| title = Vectors and Tensors in Engineering and Physics
| edition = 2/e
| year=  2003
| publisher = Westview (Perseus)
| isbn = 978-0-8133-4080-7
}}
*{{cite book
| last = Dimitrienko
| first = Yuriy
| title = Tensor Analysis and Nonlinear Tensor Functions
| year=  2002
| publisher = Kluwer Academic Publishers (Springer)
| url = http://books.google.com/books?as_isbn=140201015X
| isbn = 1-4020-1015-X
}}
*{{cite book
| last = Jeevanjee
| first = Nadir
| title = An Introduction to Tensors and Group Theory for Physicists
| year=  2011
| publisher = Birkhauser
| url = http://www.springer.com/new+%26+forthcoming+titles+(default)/book/978-0-8176-4714-8
| isbn = 978-0-8176-4714-8
}}
 
*{{cite book
| last = Lawden
| first = D. F.
| title = Introduction to Tensor Calculus, Relativity and Cosmology
| edition = 3/e
| year=  2003
| publisher = Dover
| isbn = 978-0-486-42540-5
}}
*{{cite book
| last = Lebedev
| first = Leonid P.
| coauthors = Michael J. Cloud
| title = Tensor Analysis
| year=  2003
| publisher = World Scientific
| isbn = 978-981-238-360-0
}}
*{{cite book
| last = Lovelock
| first = David
| coauthors = Hanno Rund
| title = Tensors, Differential Forms, and Variational Principles
| year=  1989
| publisher = Dover
| isbn = 978-0-486-65840-7
| origyear = 1975
}}
* Munkres, James, ''Analysis on Manifolds,'' Westview Press, 1991. Chapter six gives a "from scratch" introduction to covariant tensors.
* {{Cite journal
|title=Méthodes de calcul différentiel absolu et leurs applications
|last=Ricci
|first=Gregorio
|author-link=Gregorio Ricci-Curbastro
|last2=Levi-Civita
|first2=Tullio
|journal=Mathematische Annalen
|publisher=Springer
|volume=54
|issue=1–2
|date=March 1900
|pages=125–201
|doi=10.1007/BF01454201
|url=http://www.springerlink.com/content/u21237446l22rgg7/fulltext.pdf
|ref=harv
|postscript=<!--None-->
}}
* {{cite book |last=Kay |first=David C |title=Schaum's Outline of Tensor Calculus |publisher=McGraw-Hill |date=1988-04-01 |isbn=978-0-07-033484-7}}
* Schutz, Bernard, ''Geometrical methods of mathematical physics'', Cambridge University Press, 1980.
* {{cite book |author=Synge J.L., Schild A. |title=Tensor Calculus |publisher=first Dover Publications 1978 edition |year=  1949
|isbn=978-0-486-63612-2}}
{{Refend}}
;Specific
{{Reflist|30em}}
 
{{PlanetMath attribution|id=3112|title=tensor}}
 
==External links==
* {{Mathworld|Tensor}}
* [http://repository.tamu.edu/handle/1969.1/2502 Introduction to Vectors and Tensors, Vol 1: Linear and Multilinear Algebra] by Ray M. Bowen and C. C. Wang.
* [http://repository.tamu.edu/handle/1969.1/3609 Introduction to Vectors and Tensors, Vol 2: Vector and Tensor Analysis] by Ray M. Bowen and C. C. Wang.
* [http://www.grc.nasa.gov/WWW/K-12/Numbers/Math/documents/Tensors_TM2002211716.pdf An Introduction to Tensors for Students of Physics and Engineering] by Joseph C. Kolecki, released by [[NASA]]
* [http://nrich.maths.org/askedNRICH/edited/2604.html A discussion of the various approaches to teaching tensors, and recommendations of textbooks]
* [http://settheory.net/tensors Introduction to tensors] an original approach by S Poirier
* [http://arxiv.org/abs/math.HO/0403252 A Quick Introduction to Tensor Analysis] by R. A. Sharipov.
 
{{tensors}}
 
[[Category:Concepts in physics]]
[[Category:Tensors| ]]

Revision as of 23:59, 13 February 2014

Hello and welcome. My name is Figures Wunder. The favorite pastime for my children and me is to perform baseball but I haven't made a dime with it. Bookkeeping is her working day job now. Years ago we moved to North Dakota.

My web page: healthy meals delivered