Antenna (radio): Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>JaunJimenez
Added electronic symbol for an antenna
No edit summary
Line 1: Line 1:
{{improve|reason=remove Hermitian operators; "H is Hermitian (the name for self-adjoint in the physics literature)" and "Bounded symmetric operators are also called Hermitian." is contradicting.|date=November 2013}}
Most individuals have this habit of doing all of the stuff by themselves, irrespective of how critical or simple they are! These people won't allow others interfere in their matters. While this stance might work in additional regions of lifetime, it happens to be really not how to respond when we want to fix a Windows registry. There are some jobs like removing spywares, virus plus obsolete registry entries, that are best left to pro softwares. In this article I might tell you why it really is important to fix Windows registry NOW!<br><br>We all understand which the registry is the important component of the Windows running program because it stores all information about the Dll files, programs on the computer plus program settings. However, because days by, it is very unavoidable that we could encounter registry matter due to a huge amount of invalid, useless and unwanted entries.<br><br>Of course, the upcoming logical step is to receive these false entries cleaned out. Fortunately, this might be not a difficult task. It is the 2nd thing you need to do once you noticed your computer has lost speed. The first would be to make certain that there are no viruses or severe spyware present.<br><br>First, always clean a PC plus keep it free of dust plus dirt. Dirt clogs up all of the fans plus can cause the PC to overheat. You have to wash up disk space in order to make your computer run faster. Delete temporary and unnecessary files plus unused programs. Empty the recycle bin plus remove programs you're not using.<br><br>So to fix this, we only have to be capable to create all of the registry files non-corrupted again. This can dramatically speed up the loading time of your computer plus may allow we to a large amount of aspects on it again. And fixing these files couldn't be easier - we just have to utilize a tool called a [http://bestregistrycleanerfix.com/system-mechanic system mechanic].<br><br>Why this issue arises frequently? What are the causes of it? In fact, there are 3 major causes which could cause the PC freezing issue. To resolve the issue, we want to take 3 steps inside the following paragraphs.<br><br>Your registry is the destination all the important configurations for hardware, software and user profile configurations and needs are stored. Every time one of these items is changed, the database then starts to expand. Over time, the registry can become bloated with unnecessary files. This causes a general slow down but inside extreme situations will result significant jobs plus programs to stop functioning all together.<br><br>Before you purchase a entire new system; it's time to receive the older 1 cleaned up so you can start getting more completed online today! Visit our website under plus access the many reputable registry cleaner software accessible.
In [[mathematics]], a '''self-adjoint operator''' on a complex vector space V with [[inner product]] <math>\langle\cdot,\cdot\rangle</math> is an [[operator (mathematics)|operator]]
(a linear map A from V to itself) that is its own [[Adjoint of an operator|adjoint]]: <math>\langle Av,w\rangle=\langle v,Aw\rangle</math>.
If V is finite dimensional with a given basis, this is equivalent to the condition that the [[matrix (mathematics)|matrix]] of A is [[Hermitian matrix|Hermitian]], i.e., equal to its [[conjugate transpose]] A*.  By the finite-dimensional [[spectral theorem]], V has an [[orthonormal basis]] such that the matrix of A relative to this basis is a [[diagonal matrix]] with entries in the [[real number]]s. In this article, we consider [[generalization]]s of this [[concept]] to operators on [[Hilbert space]]s of arbitrary dimension.
 
Self-adjoint operators are used in [[functional analysis]] and [[quantum mechanics]].  In quantum mechanics their importance lies in the [[Paul Dirac|Dirac]]–[[von Neumann]] formulation of quantum mechanics, in which physical [[observable]]s such as position, [[momentum]], [[angular momentum]] and [[spin (physics)|spin]] are represented by self-adjoint operators on a Hilbert space. Of particular significance is the [[Hamiltonian (quantum mechanics)|Hamiltonian]]
 
:<math> H \psi = V \psi - \frac{\hbar^2}{2 m} \nabla^2 \psi  </math>
 
which as an observable corresponds to the total energy of a particle of mass ''m'' in a real potential field ''V''.  Differential operators are an important class of [[unbounded operator]]s.
 
The structure of self-adjoint operators on infinite-dimensional Hilbert spaces essentially resembles the
finite-dimensional case, that is to say, operators are self-adjoint if and only if they are unitarily equivalent to real-valued multiplication operators. With suitable modifications, this result can be extended to possibly unbounded operators on infinite-dimensional spaces. Since an everywhere defined self-adjoint operator is necessarily bounded, one needs be more attentive to the domain issue in the unbounded case. This is explained below in more detail.
 
==Symmetric operators ==
A [[linear operator]] ''A'' on a [[Hilbert space]] ''H'' is called '''symmetric''' if
: <math> \langle Ax \mid y \rangle =  \lang x \mid Ay \rang </math>
for all elements ''x'' and ''y'' in the domain of ''A''. 
Sometimes, such an operator is only called symmetric if it is also [[Densely defined operator|densely defined]].
 
More generally, a partially defined linear operator ''A'' from a [[topological vector space]] ''E'' into its [[continuous dual space]] ''E''<sup>∗</sup> is said to be '''symmetric''' if
:<math> \langle Ax \mid y \rangle =  \lang x \mid Ay \rang </math>
for all elements ''x'' and ''y'' in the domain of ''A''.  This usage is fairly standard in the functional analysis literature.
 
A symmetric ''everywhere defined'' operator is [[self-adjoint]].
By the [[Hellinger-Toeplitz theorem]], a symmetric ''everywhere defined'' operator is also [[operator norm|bounded]].
 
In the physics literature, the term '''Hermitian''' is used in place of the term symmetric. It should be noted, however, that the physics literature generally glosses over the distinction between operators that are merely symmetric and operators that are actually self-adjoint (as defined in the next section).
 
The previous definition agrees with the one for matrices given in the introduction to this article, if we take as ''H'' the Hilbert space '''C'''<sup>''n''</sup> with the standard dot product and interpret a square matrix as a linear operator on this Hilbert space. It is however much more general as there are important infinite-dimensional Hilbert spaces.
 
The [[spectrum of an operator|spectrum]] of any bounded symmetric operator is real; in particular all its eigenvalues are real, although a symmetric operator may have no eigenvalues.
 
A general version of the [[spectral theorem]] which also applies to bounded symmetric operators (see Reed and Simon, vol. 1, chapter VII, or other books cited) is stated below. If the set of eigenvalues for a symmetric operator is non empty, and the eigenvalues are nondegenerate, then it follows from the definition that eigenvectors corresponding to distinct eigenvalues are orthogonal.
Contrary to what is sometimes claimed in introductory physics textbooks, it is possible for symmetric operators to have no eigenvalues at all (although the [[spectrum (functional analysis)|spectrum]] of any self-adjoint operator is nonempty). The example below illustrates a special case when an (unbounded) symmetric operator does have a set of eigenvectors which constitute a Hilbert space basis. The operator ''A'' below can be seen to have a [[compact operator on Hilbert space|compact]] inverse, meaning that the corresponding differential equation ''A'' ''f'' = ''g'' is solved by some integral, therefore compact, operator ''G''. The compact symmetric operator ''G'' then has a countable family of eigenvectors which are complete in <math> L^2 </math>. The same can then be said for ''A''.
 
'''Example'''. Consider the complex Hilbert space L<sup>2</sup>[0,1] and the [[differential operator]]
 
: <math> A = - \frac{d^2}{dx^2} </math>
 
defined on the subspace consisting of all complex-valued infinitely [[differentiable]] functions ''f'' on  [0,1] with the boundary conditions
 
:<math> f(0) = f(1) = 0 \quad </math>
 
Then [[integration by parts]] shows that ''A'' is symmetric.  Its eigenfunctions are the sinusoids
 
:<math> f_n(x) =  \sin(n \pi x) \quad  n= 1,2, \ldots </math>
 
with the real eigenvalues ''n''<sup>2</sup>π<sup>2</sup>; the well-known orthogonality of the sine functions follows as a consequence of the property of being symmetric.
 
We consider generalizations of this operator below.
 
== Self-adjoint operators ==
 
Given a densely defined linear operator ''A'' on ''H'', its adjoint ''A''* is defined as follows:
* The domain of ''A''* consists of vectors ''x'' in ''H'' such that
 
::<math> y \mapsto \langle x \mid A y \rangle </math>
 
: (which is a densely defined ''linear'' map) is a continuous linear functional.  By continuity and density of the domain of ''A'',  it extends to a unique continuous linear functional on all of ''H''.
 
* By the [[Riesz representation theorem]] for linear functionals, if ''x'' is in the domain of ''A''*, there is a unique vector ''z'' in ''H'' such that
::<math> \langle x  \mid A y \rangle  = \langle z  \mid  y \rangle \quad \forall y \in \operatorname{dom} A </math>
:This  vector ''z'' is defined to be ''A''* ''x''.  It can be shown that the dependence of ''z'' on ''x'' is linear.
 
Notice that it is the denseness of the domain of the operator, along with the uniqueness part of Riesz representation, that ensures the adjoint operator is well defined.
 
A result of Hellinger-Toeplitz type says that an operator having an everywhere defined bounded adjoint is bounded.
 
The condition for a linear operator on a Hilbert space to be ''self-adjoint'' is stronger than to be ''symmetric''. Although this distinction is technical, it is very important; the spectral theorem applies only to operators that are self-adjoint and not to operators that are merely symmetric. For an extensive discussion of the distinction, see Chapter 9 of Hall (2013).
 
For any densely defined operator ''A'' on Hilbert space one can define its adjoint operator ''A''*.
For a symmetric operator ''A'', the domain of the operator ''A''* contains the domain of the operator ''A'', and the restriction of the operator ''A''* on the domain of ''A'' coincides with the operator ''A'', i.e. <math> A \subseteq A^* </math>, in other words ''A''* is extension of ''A''. For a self-adjoint operator ''A'' the domain of ''A''* is the same as the domain of ''A'', and ''A''=''A''*. See also [[Extensions of symmetric operators]] and [[unbounded operator]].
 
===Geometric interpretation===
There is a useful [[geometric]]al way of looking at the adjoint of an operator ''A'' on ''H'' as follows: we consider the graph G(''A'') of ''A'' defined by
 
:<math> \operatorname{G}(A) = \{(\xi, A \xi): \xi \in \operatorname{dom}(A)\} \subseteq H \oplus H .</math>
 
'''Theorem'''.  Let J be the [[symplectic matrix|symplectic mapping]]
 
: <math> H \oplus H \rightarrow H \oplus H </math>
 
given by
 
:<math> \operatorname{J}: (\xi, \eta) \mapsto (-\eta, \xi). </math>
 
Then the graph of ''A''* is the [[orthogonal complement]] of JG(''A''):
 
:<math> \operatorname{G}(A^*) = (\operatorname{J}\operatorname{G}(A))^\perp
= \{
(x,y) \in H \oplus H : \langle (x,y)|(-A\xi,\xi) \rangle = 0
\;\;\forall  \xi \in \operatorname{dom}(A)\}
</math>
 
A densely defined operator ''A'' is symmetric [[if and only if]]
:<math> A \subseteq A^*.</math>
 
where the subset notation <math>A \subseteq A^*</math> is understood to mean <math> G(A) \subseteq G(A^*).</math> An operator ''A'' is '''self-adjoint''' if and only if <math>A=A^*</math>; that is, if and only if <math>G(A)=G(A^*).</math>
 
'''Example'''.  Consider the complex Hilbert space L<sup>2</sup>('''R'''), and the operator which multiplies a given function by ''x'':
 
:<math> A f(x) = xf(x) </math>
 
The domain of ''A'' is  the space of all L<sup>2</sup> functions for which the right-hand-side is square-integrable. ''A'' is a symmetric operator without any eigenvalues and eigenfunctions.  In fact it turns out that the operator is self-adjoint, as follows from the theory outlined below.
 
As we will see later, self-adjoint operators have very important spectral properties; they are in fact multiplication operators on general measure spaces.
 
== Spectral theorem ==
{{Main|Spectral theorem}}
 
Partially defined operators ''A'', ''B'' on Hilbert spaces ''H'', ''K'' are '''unitarily equivalent''' if and only if there is a [[unitary transformation]] ''U'':''H'' → ''K'' such that
 
* ''U'' maps dom ''A'' [[bijective]]ly onto  dom ''B'',
 
* <math> B U \xi = U A \xi ,\quad \xi \in \operatorname{dom}A. </math>
 
A [[multiplication operator]] is defined as follows:  Let <math> (X, \Sigma, \mu) </math> be a countably additive [[measure space]] and ''f'' a real-valued measurable function on ''X''. An operator ''T'' of the form
 
:<math> [T \psi] (x) = f(x) \psi(x) \quad </math>
 
whose domain is the space of ψ for which the right-hand side above is in ''L''<sup>2</sup> is called a multiplication operator.
 
'''Theorem'''. Any  multiplication operator is a (densely defined) self-adjoint operator. Any self-adjoint operator is unitarily equivalent to a multiplication operator.
 
This version of the spectral theorem for self-adjoint operators can be proved by reduction to the spectral theorem for unitary operators.  This reduction uses the ''Cayley transform'' for self-adjoint operators which is defined in the next section. We might note that if T is multiplication by f, then the spectrum of T is just the [[essential range]] of f.
 
=== Borel functional calculus ===
 
Given the representation of ''T'' as a multiplication operator, it is easy to characterize the '''[[Borel functional calculus]]''':  If ''h'' is a bounded real-valued Borel function on '''R''', then ''h''(''T'') is the operator of multiplication by the composition <math> h \circ f </math>.  In order for this to be well-defined, we must show that it is the unique operation on  bounded real-valued Borel functions satisfying a number of conditions.
 
=== Resolution of the identity ===
 
It has been customary to introduce the following notation
 
:<math> \operatorname{E}_T(\lambda) = \mathbf{1}_{(-\infty, \lambda]} (T) </math>
 
where <math>\mathbf{1}_{(-\infty, \lambda]}</math> denotes the function which is identically 1 on the interval <math>(-\infty, \lambda]</math>.  The family of projection operators E<sub>''T''</sub>(λ) is called '''resolution of the identity''' for ''T''.  Moreover, the following [[Stieltjes integral]] representation for ''T'' can be proved:
 
:<math> T = \int_{-\infty}^{+\infty} \lambda d \operatorname{E}_T(\lambda). </math>
 
The definition of the operator integral above can be reduced to that that of a scalar valued Stieltjes integral using the weak operator topology. In more modern treatments however, this representation is usually avoided, since most technical problems can be dealt with by the functional calculus.
 
=== Formulation in the physics literature ===
 
In physics, particularly in quantum mechanics, the spectral theorem is expressed in a way which combines the spectral theorem as stated above and the [[Borel functional calculus]] using [[Dirac notation]] as follows:
 
If ''H'' is self-adjoint and ''f'' is a [[Borel function]],
 
:<math>
f(H)= \int dE |\Psi_{E}\rangle f(E) \langle \Psi_{E} |
</math>
 
with
 
:<math>
H | \Psi_{E}\rangle = E |\Psi_{E}\rangle
</math>
 
where the integral runs over the whole spectrum of ''H''. The notation suggests that ''H'' is diagonalized by the eigenvectors Ψ<sub>''E''</sub>. Such a notation is purely [[Formal calculation|formal]]. One can see the similarity between Dirac's notation and the previous section. The resolution of the identity (sometimes called projection valued measures) formally resembles the rank-1 projections <math> | \Psi_{E}\rangle \langle \Psi_{E} |</math>.
In the Dirac notation, (projective) measurements are described via [[eigenvalues]] and [[eigenstates]], both purely formal objects. As one would expect, this does not survive passage to the resolution of the identity. In the latter formulation, measurements are described using the [[spectral measure]] of <math> | \Psi \rangle </math>, if the system is prepared in <math> | \Psi \rangle </math> prior to the measurement. Alternatively, if one would like to preserve the notion of eigenstates and make it rigorous, rather than merely formal, one can replace the state space by a suitable [[rigged Hilbert space]].
 
If {{nowrap|''f'' {{=}} 1}}, the theorem is referred to as resolution of unity:
 
:<math>
I = \int dE | \Psi_{E}\rangle \langle \Psi_{E} |
</math>
 
In the case <math>H_{\text{eff}}=H-i\Gamma</math> is the sum of an Hermitian ''H'' and a skew-Hermitian  (see [[skew-Hermitian matrix]]) operator <math> -i\Gamma</math>, one defines the [[biorthogonal system|biorthogonal]] basis set
 
:<math> H^*_{\text{eff}} | \Psi_{E}^* \rangle = E^* | \Psi_{E}^*\rangle</math>
 
and write the spectral theorem as:
 
:<math>
f(H_{\text{eff}})= \int dE | \Psi_{E}\rangle f(E) \langle \Psi_{E}^* |
</math>
 
(See [[Feshbach–Fano partitioning]] method for the context where such operators appear in [[scattering theory]]).
 
== Extensions of symmetric operators ==
 
{{Further2|[[Extensions of symmetric operators]] and [[unbounded operator]]}}
 
The following question arises in several contexts: if an operator ''A'' on the Hilbert space ''H'' is symmetric, when does it have self-adjoint extensions?  One answer is provided by the '''[[Cayley transform]]''' of a self-adjoint operator and the deficiency indices. (We should note here that it is often of technical convenience to deal with [[closed operator]]s. In the symmetric case, the closedness requirement poses no obstacles, since it is known that all symmetric operators are [[closable operator|closable]].)
 
'''Theorem'''. Suppose ''A'' is a symmetric operator.  Then there is a
unique partially defined  linear operator
 
: <math>\operatorname{W}(A) \colon \operatorname{ran}(A+i) \rightarrow \operatorname{ran}(A-i)</math>
 
such that
 
: <math> \operatorname{W}(A)(Ax + ix) = Ax - ix \quad  x \in \operatorname{dom}(A). </math>
 
Here, ''ran'' and ''dom'' denote the [[range (mathematics)|range]] and the [[domain (mathematics)|domain]], respectively. W(''A'') is [[isometry|isometric]] on its domain.  Moreover, the range of 1&nbsp;−&nbsp;W(''A'') is [[dense set|dense]] in ''H''.
 
Conversely, given any partially defined operator ''U'' which is isometric on its domain (which is not
necessarily closed) and such that 1&nbsp;−&nbsp;''U'' is dense, there is a (unique)  operator S(''U'')
 
: <math>\operatorname{S}(U) \colon \operatorname{ran}(1 - U) \rightarrow \operatorname{ran}(1+U)</math>
 
such that
 
: <math> \operatorname{S}(U)(x - Ux)= i(x + U x) \quad x \in \operatorname{dom}(U). </math>
 
The operator S(''U'')  is densely defined and symmetric.
 
The mappings W and S are inverses of each other.
 
The mapping W is called the '''Cayley transform'''.  It associates a [[partial isometry|partially defined isometry]] to any symmetric densely defined operator.  Note that the mappings W and S are [[monotone convergence theorem|monotone]]: This means that if ''B'' is a symmetric operator that extends the densely defined symmetric operator ''A'', then W(''B'') extends W(''A''), and similarly for S.
 
'''Theorem'''. A necessary and sufficient condition for ''A'' to be self-adjoint is that its Cayley transform W(''A'') be unitary.
 
This immediately gives us a necessary and sufficient condition for ''A'' to have a self-adjoint extension, as follows:
 
'''Theorem'''. A necessary and sufficient condition for ''A'' to have a self-adjoint extension is that W(''A'') have a unitary extension.
 
A partially defined isometric operator ''V'' on a Hilbert space ''H'' has a unique isometric extension to the norm closure of dom(''V''). A partially defined isometric operator with closed domain is called a [[partial isometry]].
 
Given a partial isometry ''V'', the '''deficiency indices''' of ''V'' are defined as the dimension of the [[orthogonal complement]]s of the domain and range:
 
:<math> n_+(V) = \operatorname{dim}\ \operatorname{dom}(V)^{\perp}</math>
 
:<math> n_-(V) = \operatorname{dim} \ \operatorname{ran}(V)^{\perp}</math>
 
'''Theorem'''. A partial isometry ''V'' has a unitary extension if and only if the deficiency indices are identical. Moreover, ''V'' has a ''unique'' unitary extension if and only if the both deficiency indices are zero.
 
We see that there is a bijection between symmetric extensions of an operator and isometric extensions of its Cayley transform. An operator which has a unique self-adjoint extension is said to be '''essentially self-adjoint'''. Such operators have a well-defined [[Borel functional calculus]]. Symmetric operators which are not essentially self-adjoint may still have a [[canonical form|canonical]] self-adjoint extension.  Such is the case for ''non-negative'' symmetric operators (or more generally, operators which are bounded below).  These operators always have a canonically defined [[Friedrichs extension]] and for these operators we can define a canonical functional calculus. Many operators that occur in  analysis are bounded below (such as the negative of the [[Laplacian]] operator), so the issue of essential adjointness for these operators is less critical.
 
===Self-adjoint extensions in quantum mechanics ===
In quantum mechanics, observables correspond to self-adjoint operators. By [[Stone's theorem on one-parameter unitary groups]], self-adjoint operators are precisely the infinitesimal generators of unitary groups of [[time evolution]] operators. However, many physical problems are formulated as a time-evolution equation involving differential operators for which the Hamiltonian is only symmetric. In such cases, either the Hamiltonian is essentially self-adjoint, in which case the physical problem has unique solutions or one attempts to find self-adjoint extensions of the Hamiltonian corresponding to different types of boundary conditions or conditions at infinity.
 
'''Example.''' The one-dimensional Schrödinger operator with the potential <math> V(x) = -(1+|x|)^\alpha</math>, defined initially on smooth compactly supported functions, is essentially self-adjoint (that is, has a self-adjoint closure) for <math>0 < \alpha \le 2</math> but not for <math>\alpha > 2.</math> See Berezin and Schubin, pages 55 and 86, or Section 9.10 in Hall.
 
'''Example.''' There is no self-adjoint momentum operator <math>p</math> for a particle moving on a half-line. Nevertheless, the Hamiltonian <math>p^2</math> of a "free" particle on a half-line has several self-adjoint extensions corresponding to different types of boundary conditions. Physically, these boundary conditions are related to reflections of the particle at the origin (see Reed and Simon, vol.2).
 
== Von Neumann's formulas ==
 
Suppose ''A'' is symmetric densely defined. Then any symmetric extension of ''A'' is a restriction of ''A''*.  Indeed, if ''B'' is symmetric, then
 
:<math> A \subseteq B \implies B \subseteq B^* \subseteq A^* </math>.
 
'''Theorem'''.  Suppose ''A'' is a densely defined symmetric operator. Let
 
:<math> N_{+} = \operatorname{ran}(A + i)^{\perp}, </math>
 
:<math> N_{-} = \operatorname{ran}(A - i)^{\perp} </math>.
 
Then
 
:<math> N_{+} = \operatorname{ker}(A^* - i) </math> and
 
:<math> N_{-} = \operatorname{ker}(A^* + i) </math>.
 
and
 
:<math> \operatorname{dom}(A^*) = \operatorname{dom}(\overline{A}) \oplus N_{+} \oplus N_{-} </math>,
 
where the decomposition is orthogonal relative to the graph inner product of dom(''A''*):
 
:<math> \langle \xi | \eta \rangle_{\mathrm{graph}} = \langle \xi | \eta \rangle + \langle A^* \xi | A^* \eta \rangle </math>.
 
These are referred to as von Neumann's formulas in the Akhiezer and Glazman reference.
 
== Examples ==
 
We first consider the differential operator
 
: <math> D: \phi \mapsto \frac{1}{i} \phi' </math>
 
defined on the space of complex-valued C<sup>∞</sup> functions on [0,1] vanishing near 0 and 1.  ''D'' is a symmetric operator as can be shown by [[integration by parts]].  The spaces ''N''<sub>+</sub>, ''N''<sub>&minus;</sub> are given respectively by the [[distribution (mathematics)|distribution]]al solutions to the equation
 
:<math> u' = i u \quad </math>
 
:<math> u' =  - i u \quad </math>
 
which are in ''L''<sup>2</sup> [0,1].  One can show that each one of these solution spaces is 1-dimensional, generated by the functions
''x'' → ''e''<sup>''ix''</sup> and ''x'' → ''e''<sup>−''ix''</sup> respectively. This shows that ''D'' is not essentially self-adjoint, but does have self-adjoint extensions.  These self-adjoint extensions are parametrized by the space of unitary mappings
 
: <math> N_{+} \rightarrow N_{-} \,\!</math>
 
which in this case happens to be the unit circle '''T'''.
 
This simple example illustrates a general fact about self-adjoint extensions of symmetric differential operators ''P'' on an open set ''M''. They are determined by the unitary maps between the eigenvalue spaces
 
: <math> N_\pm = \{u \in L^2(M): P_{\operatorname{dist}} u =  \pm i u\} </math>
 
where ''P''<sub>dist</sub> is the distributional extension of ''P''.
 
We next give the example of differential operators with [[constant coefficient]]s. Let
 
:<math> P(\vec{x}) = \sum_\alpha c_\alpha x^\alpha </math>
 
be a polynomial on '''R'''<sup>''n''</sup> with ''real'' coefficients, where α ranges over a (finite) set of [[multi-index|multi-indices]].  Thus
 
: <math> \alpha = (\alpha_1, \alpha_2, \ldots, \alpha_n) \,\!</math>
 
and
 
: <math> x^\alpha = x_1^{\alpha_1} x_2^{\alpha_2} \cdots x_n^{\alpha_n}. </math>
 
We also use the notation
 
:<math> D^\alpha = \frac{1}{i^{|\alpha|}} \partial_{x_1}^{\alpha_1}\partial_{x_2}^{\alpha_2}  \cdots \partial_{x_n}^{\alpha_n}. </math>
 
Then the operator ''P''(D) defined on the space of infinitely differentiable functions of compact support on '''R'''<sup>''n''</sup> by
 
:<math> P(\operatorname{D}) \phi = \sum_\alpha c_\alpha D^\alpha \phi </math>
 
is essentially self-adjoint on ''L''<sup>2</sup>('''R'''<sup>''n''</sup>).
 
'''Theorem'''. Let ''P'' a polynomial function on '''R'''<sup>''n''</sup> with real coefficients, '''F''' the Fourier transform considered as a unitary map ''L''<sup>2</sup>('''R'''<sup>''n''</sup>) → ''L''<sup>2</sup>('''R'''<sup>''n''</sup>).  Then '''F'''* ''P''(D) '''F''' is essentially self-adjoint and its unique self-adjoint extension is the operator of multiplication by the function ''P''.
 
More generally, consider linear differential operators acting on infinitely differentiable complex-valued functions of compact support.  If ''M'' is an open subset of '''R'''<sup>''n''</sup>
 
:<math> P \phi(x) = \sum_\alpha a_\alpha (x) [D^\alpha \phi](x) \quad </math>
 
where ''a''<sub>α</sub> are (not necessarily constant) infinitely differentiable functions. ''P'' is a linear operator
 
:<math> C_0^\infty(M) \rightarrow C_0^\infty(M). \quad </math>
 
Corresponding to ''P'' there is another differential operator, the '''[[formal adjoint]]''' of ''P''
 
:<math> P^{\mathrm{*form}} \phi = \sum_\alpha D^\alpha (\overline{a_\alpha} \phi) \quad </math>
 
'''Theorem'''.  The operator theoretic adjoint ''P''* of ''P'' is a restriction of the distributional extension of the formal adjoint. Specifically:
 
:<math> \operatorname{dom} P^* = \{u \in L^2(M): P^{\mathrm{*form}}u
\in L^2(M)\}. </math>
 
== Spectral multiplicity theory ==
 
The multiplication representation of a self-adjoint operator, though extremely useful, is not a canonical representation.  This suggests that it is not easy to extract from this representation a criterion to determine when self-adjoint operators ''A'' and ''B'' are unitarily equivalent.  The finest grained representation which we now discuss involves spectral multiplicity.  This circle of results is called the ''[[Hans Hahn (mathematician)|Hahn]]-[[Ernst Hellinger|Hellinger]] theory of spectral multiplicity''.
 
We first define ''uniform multiplicity'':
 
'''Definition'''. A self-adjoint operator ''A'' has uniform multiplicity ''n'' where ''n'' is such that 1 ≤ ''n'' ≤ ω
if and only if ''A'' is unitarily equivalent  to the operator M<sub>''f''</sub> of multiplication by the function ''f''(λ) = λ on
 
:<math> L^2_{\mu}(\mathbb{R}, \mathbf{H}_n)= \{\psi: \mathbb{R} \rightarrow \mathbf{H}_n: \psi \mbox{ measurable and } \int_{\mathbb{R}} \|\psi(t)\|^2 d \mu(t) < \infty\} </math>
 
where  '''H'''<sub>''n''</sub> is a Hilbert space of dimension ''n''.  The domain of  M<sub>''f''</sub> consists of vector-valued functions ψ on '''R''' such that
 
:<math> \int_{\mathbb{R}} |\lambda|^2 \ \| \psi(\lambda)\|^2 \, d \mu(\lambda) < \infty. </math>
 
Non-negative countably additive measures μ, ν are '''mutually singular''' if and only if they are supported on disjoint Borel sets.
 
'''Theorem'''.  Let ''A'' be a self-adjoint operator on a ''separable'' Hilbert space ''H''.  Then there is an ω sequence of countably additive finite measures on '''R''' (some of which may be identically 0)
 
:<math> \{\mu_\ell\}_{1 \leq \ell \leq \omega} </math>
 
such that the measures are pairwise singular and ''A'' is unitarily equivalent to the operator of multiplication by the function ''f''(λ) = λ on
 
:<math> \bigoplus_{1 \leq \ell \leq \omega} L^2_{\mu_\ell}(\mathbb{R}, \mathbf{H}_\ell). </math>
 
This representation is unique in the following sense: For any two such representations of the same ''A'', the corresponding measures are equivalent in the sense that they have the same sets of measure 0.
 
The spectral multiplicity theorem can be reformulated using the language of [[direct integral]]s of Hilbert spaces:
 
'''Theorem'''.  Any self-adjoint operator on a separable Hilbert space is unitarily equivalent to multiplication by the function λ → λ on
 
:<math> \int_\mathbb{R}^\oplus H_x d \mu(x). </math>
 
The measure equivalence class of μ (or equivalently its sets of measure 0) is uniquely determined and the measurable family
{''H''<sub>''x''</sub>}<sub>''x''</sub> is determined almost everywhere with respect to μ.
 
=== Example: structure of the Laplacian ===
The Laplacian on '''R'''<sup>''n''</sup> is the operator
 
:<math> \Delta = \sum_{i=1}^n \partial_{x_i}^2. </math>
 
As remarked above, the Laplacian is diagonalized by the Fourier transform.  Actually it is more natural to consider the ''negative'' of the Laplacian - Δ since as an operator it is non-negative; (see [[elliptic operator]]).
 
'''Theorem'''.  If ''n''=1, then -Δ has uniform multiplicity '''mult'''=2, otherwise -Δ has uniform multiplicity '''mult'''=ω.  Moreover, the measure μ<sub>'''mult'''</sub> is Borel measure on <nowiki>[</nowiki>0, ∞).
 
== Pure point spectrum ==
A self-adjoint operator ''A'' on ''H'' has pure point spectrum if and only if  ''H'' has an orthonormal basis {''e''<sub>''i''</sub>}<sub>''i'' ∈ I</sub> consisting of eigenvectors for ''A''.
 
'''Example'''. The Hamiltonian for the harmonic oscillator has a quadratic potential ''V'', that is
 
:<math> -\Delta  + |x|^2. \quad </math>
 
This Hamiltonian has pure point spectrum; this is typical for bound state [[Hamiltonian (quantum mechanics)|Hamiltonians]] in quantum mechanics. As was pointed out in a previous example, a sufficient condition that an unbounded symmetric operator has eigenvectors which form a Hilbert space basis is that it has a compact inverse.
 
== See also ==
*[[Compact operator on Hilbert space]]
*[[Theoretical and experimental justification for the Schrödinger equation]]
*[[Unbounded operator]]
 
== References ==
*{{cite book |authorlink=Naum Akhiezer |first=N. I. |last=Akhiezer |first2=I. M. |last2=Glazman |title=Theory of Linear Operators in Hilbert Space |others=Two volumes |location= |publisher=Pitman |year=1981 |isbn= }}
*{{cite book |first=F. A. |last=Berezin |first2=M. A. |last2=Shubin |title=The Schrödinger Equation |location= |publisher=Kluwer |year=1991 }}
*{{cite book |authorlink=Tosio Kato |first=T. |last=Kato |title=Perturbation Theory for Linear Operators |publisher=Springer |location=New York |year=1966 |isbn= }}
*{{cite book |first=B. C. |last=Hall |title=Quantum Theory for Mathematicians |publisher=Springer |location=New York |year=2013 |isbn= }}
*{{cite book |authorlink=Michael C. Reed |first=M. |last=Reed |authorlink2=Barry Simon |first2=B. |last2=Simon |title=Methods of Mathematical Physics |others=Vol 2 |publisher=Academic Press |year=1972 |isbn= }}
*{{cite book |authorlink=Gerald Teschl |first=G. |last=Teschl |title=Mathematical Methods in Quantum Mechanics; With Applications to Schrödinger Operators |publisher=American Mathematical Society |location=Providence |year=2009 |url=http://www.mat.univie.ac.at/~gerald/ftp/book-schroe/ }}
*{{cite book |authorlink=Kōsaku Yosida |first=K. |last=Yosida |title=Functional Analysis |location= |publisher=Academic Press |year=1965 |isbn= }}
 
{{Functional Analysis}}
 
{{DEFAULTSORT:Self-Adjoint Operator}}
[[Category:Operator theory]]
[[Category:Hilbert space]]

Revision as of 16:03, 27 February 2014

Most individuals have this habit of doing all of the stuff by themselves, irrespective of how critical or simple they are! These people won't allow others interfere in their matters. While this stance might work in additional regions of lifetime, it happens to be really not how to respond when we want to fix a Windows registry. There are some jobs like removing spywares, virus plus obsolete registry entries, that are best left to pro softwares. In this article I might tell you why it really is important to fix Windows registry NOW!

We all understand which the registry is the important component of the Windows running program because it stores all information about the Dll files, programs on the computer plus program settings. However, because days by, it is very unavoidable that we could encounter registry matter due to a huge amount of invalid, useless and unwanted entries.

Of course, the upcoming logical step is to receive these false entries cleaned out. Fortunately, this might be not a difficult task. It is the 2nd thing you need to do once you noticed your computer has lost speed. The first would be to make certain that there are no viruses or severe spyware present.

First, always clean a PC plus keep it free of dust plus dirt. Dirt clogs up all of the fans plus can cause the PC to overheat. You have to wash up disk space in order to make your computer run faster. Delete temporary and unnecessary files plus unused programs. Empty the recycle bin plus remove programs you're not using.

So to fix this, we only have to be capable to create all of the registry files non-corrupted again. This can dramatically speed up the loading time of your computer plus may allow we to a large amount of aspects on it again. And fixing these files couldn't be easier - we just have to utilize a tool called a system mechanic.

Why this issue arises frequently? What are the causes of it? In fact, there are 3 major causes which could cause the PC freezing issue. To resolve the issue, we want to take 3 steps inside the following paragraphs.

Your registry is the destination all the important configurations for hardware, software and user profile configurations and needs are stored. Every time one of these items is changed, the database then starts to expand. Over time, the registry can become bloated with unnecessary files. This causes a general slow down but inside extreme situations will result significant jobs plus programs to stop functioning all together.

Before you purchase a entire new system; it's time to receive the older 1 cleaned up so you can start getting more completed online today! Visit our website under plus access the many reputable registry cleaner software accessible.