Conductor of an abelian variety: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>Citation bot 1
m [Pu344]+: jstor.
 
en>Uni.Liu
Line 1: Line 1:
Bonus: WP Twin and WP Twin Auto Backup: (link to  ) While not a theme, I think this software is essential if you are maintaining your Wordpress blog or regularly create new blog sites. If you're ready to find more info about [http://GET7.pw/wordpress_backup_729796 wordpress dropbox backup] check out our own website. What I advise you do next is save the backup data file to a remote place like a CD-ROM, external disk drive if you have one or a provider such as Dropbox. The effect is to promote older posts by moving them back onto the front page and into the rss feed. If you're using Wordpress and want to make your blog a "dofollow" blog, meaning that links from your blog pass on the benefits of Google pagerank, you can install one of the many dofollow plugins available. provided by Word - Press Automatic Upgrade, so whenever you need to update the new version does not, it automatically creates no webmaster. <br><br>Word - Press is known as the most popular blogging platform all over the web and is used by millions of blog enthusiasts worldwide. If you are a positive thinker businessman then today you have to put your business online. Some plugins ask users to match pictures or add numbers, and although effective, they appear unprofessional and unnecessary. Being able to help with your customers can make a change in how a great deal work, repeat online business, and referrals you'll be given. As soon as you start developing your Word - Press MLM website you'll see how straightforward and simple it is to create an online presence for you and the products and services you offer. <br><br>Saying that, despite the launch of Wordpress Express many months ago, there has still been no sign of a Wordpress video tutorial on offer UNTIL NOW. Browse through the popular Wordpress theme clubs like the Elegant Themes, Studio Press, Woo - Themes, Rocket Theme, Simple Themes and many more. Use this section to change many formatting elements. Newer programs allow website owners and internet marketers to automatically and dynamically change words in their content to match the keywords entered by their web visitors in their search queries'a feat that they cannot easily achieve with older software. If you've hosted your Word - Press website on a shared hosting server then it'll be easier for you to confirm the restricted access to your site files. <br><br>Word - Press has plenty of SEO benefits over Joomla and Drupal. The SEOPressor Word - Press SEO Plugin works by analysing each page and post against your chosen keyword (or keyword phrase) and giving a score, with instructions on how to improve it. Exacting subjects in reality must be accumulated in head ahead of planning on your high quality theme. Giant business organizations can bank on enterprise solutions to incorporate latest web technologies such as content management system etc, yet some are looking for economical solutions. Wordpress template is loaded with lots of prototype that unite graphic features and content area. <br><br>More it extends numerous opportunities where your firm is at comfort and rest assured of no risks & errors. An ease of use which pertains to both internet site back-end and front-end users alike. By the time you get the Gallery Word - Press Themes, the first thing that you should know is on how to install it. If this is not possible you still have the choice of the default theme that is Word - Press 3. Get started today so that people searching for your type of business will be directed to you.
In [[probability theory]], and in particular, [[information theory]], the '''conditional mutual information''' is, in its most basic form, the [[expected value]] of the [[mutual information]] of two random variables given the value of a third.
 
==Definition==
For discrete random variables <math>X,</math> <math>Y,</math> and <math>Z,</math> we define
:<math>I(X;Y|Z) = \mathbb E_Z \big(I(X;Y)|Z\big)
    = \sum_{z\in Z} p_Z(z) \sum_{y\in Y} \sum_{x\in X}
      p_{X,Y|Z}(x,y|z) \log \frac{p_{X,Y|Z}(x,y|z)}{p_{X|Z}(x|z)p_{Y|Z}(y|z)},</math>
where the marginal, joint, and/or conditional [[probability mass function]]s are denoted by <math>p</math> with the appropriate subscript. This can be simplified as
:<math>I(X;Y|Z) = \sum_{z\in Z} \sum_{y\in Y} \sum_{x\in X}
      p_{X,Y,Z}(x,y,z) \log \frac{p_Z(z)p_{X,Y,Z}(x,y,z)}{p_{X,Z}(x,z)p_{Y,Z}(y,z)}.</math>
Alternatively, we may write<ref>K. Makarychev et al. ''A new class of non-Shannon-type inequalities for entropies.'' Communications in Information and Systems, Vol. 2, No. 2, pp. 147&ndash;166, December 2002 [http://www.cs.princeton.edu/~ymakaryc/papers/nonshann.pdf PDF]</ref>
:<math>I(X;Y|Z) = H(X,Z) + H(Y,Z) - H(X,Y,Z) - H(Z)
                = H(X|Z) - H(X|Y,Z)</math>
This can be rewritten to show its relationship to mutual information
:<math>I(X;Y|Z) = I(X;Y,Z) - I(X;Z)</math>
usually rearranged as '''the chain rule for mutual information'''
:<math>I(X;Y,Z) = I(X;Z) + I(X;Y|Z)</math>
Another equivalent form of the above is
:<math>I(X;Y|Z) = H(Z|X) + H(X) + H(Z|Y) + H(Y) - H(Z|X,Y) - H(X,Y) - H(Z)
                = I(X;Y) + H(Z|X) + H(Z|Y) - H(Z|X,Y) - H(Z)</math>
 
Conditioning on a third random variable may either increase or decrease the mutual information: that is, the difference <math>I(X;Y|Z) - I(X;Y)</math>, called the [[interaction information]], may be positive, negative, or zero, but it is always true that
:<math>I(X;Y|Z) \ge 0</math>
for discrete, jointly distributed random variables ''X'', ''Y'', ''Z''.  This result has been used as a basic building block for proving other [[inequalities in information theory]], in particular, those known as Shannon-type inequalities.
 
Like mutual information, conditional mutual information can be expressed as a [[Kullback-Leibler divergence]]:
 
:<math> I(X;Y|Z) = D_{\mathrm{KL}}[ p(X,Y,Z) \| p(X|Z)p(Y|Z)p(Z) ]. </math>
 
Or as an expected value of simpler Kullback-Leibler divergences:
:<math> I(X;Y|Z) = \sum_{z \in Z} p( Z=z ) D_{\mathrm{KL}}[ p(X,Y|z) \| p(X|z)p(Y|z) ], </math>
:<math> I(X;Y|Z) = \sum_{y \in Y} p( Y=y ) D_{\mathrm{KL}}[ p(X,Z|y) \| p(X|Z)p(Z|y) ]. </math>
 
==More general definition==
A more general definition of conditional mutual information, applicable to random variables with continuous or other arbitrary distributions, will depend on the concept of '''[[regular conditional probability]]'''.  (See also.<ref>[http://planetmath.org/encyclopedia/ConditionalProbabilityMeasure.html Regular Conditional Probability] on [http://planetmath.org/ PlanetMath]</ref><ref>D. Leao Jr. et al. ''Regular conditional probability, disintegration of probability and Radon spaces.'' Proyecciones. Vol. 23, No. 1, pp. 15–29, May 2004, Universidad Católica del Norte, Antofagasta, Chile [http://www.scielo.cl/pdf/proy/v23n1/art02.pdf PDF]</ref>)
 
Let <math>(\Omega, \mathcal F, \mathfrak P)</math> be a [[probability space]], and let the random variables ''X'', ''Y'', and ''Z'' each be defined as a Borel-measurable function from <math>\Omega</math> to some state space endowed with a topological structure.
 
Consider the Borel measure (on the σ-algebra generated by the open sets) in the state space of each random variable defined by assigning each Borel set the <math>\mathfrak P</math>-measure of its preimage in <math>\mathcal F</math>.  This is called the [[pushforward measure]] <math>X _* \mathfrak P = \mathfrak P\big(X^{-1}(\cdot)\big).</math>  The '''support of a random variable''' is defined to be the [[Support (measure theory)|topological support]] of this measure, i.e. <math>\mathrm{supp}\,X = \mathrm{supp}\,X _* \mathfrak P.</math>
 
Now we can formally define the [[conditional probability distribution|conditional probability measure]] given the value of one (or, via the [[product topology]], more) of the random variables. Let <math>M</math> be a measurable subset of <math>\Omega,</math> (i.e. <math>M \in \mathcal F,</math>) and let <math>x \in \mathrm{supp}\,X.</math>  Then, using the [[disintegration theorem]]:
:<math>\mathfrak P(M | X=x) = \lim_{U \ni x}
  \frac {\mathfrak P(M \cap \{X \in U\})}
        {\mathfrak P(\{X \in U\})}
  \qquad \textrm{and} \qquad \mathfrak P(M|X) = \int_M d\mathfrak P\big(\omega|X=X(\omega)\big),</math>
where the limit is taken over the open neighborhoods <math>U</math> of <math>x</math>, as they are allowed to become arbitrarily smaller with respect to [[Subset|set inclusion]]. 
 
Finally we can define the conditional mutual information via [[Lebesgue integration]]:
:<math>I(X;Y|Z) = \int_\Omega \log
  \frac {d \mathfrak P(\omega|X,Z)\, d\mathfrak P(\omega|Y,Z)}
        {d \mathfrak P(\omega|Z)\, d\mathfrak P(\omega|X,Y,Z)}
  d \mathfrak P(\omega),
  </math>
where the integrand is the logarithm of a [[Radon–Nikodym derivative]] involving some of the conditional probability measures we have just defined.
 
==Note on notation==
In an expression such as <math>I(A;B|C),</math> <math>A,</math> <math>B,</math> and <math>C</math> need not necessarily be restricted to representing individual random variables, but could also represent the joint distribution of any collection of random variables defined on the same [[probability space]]. As is common in [[probability theory]], we may use the comma to denote such a joint distribution, e.g. <math>I(A_0,A_1;B_1,B_2,B_3|C_0,C_1).</math>  Hence the use of the semicolon (or occasionally a colon or even a wedge <math>\wedge</math>) to separate the principal arguments of the mutual information symbol.  (No such distinction is necessary in the symbol for [[joint entropy]], since the joint entropy of any number of random variables is the same as the entropy of their joint distribution.)
 
==Multivariate mutual information==
{{main|Multivariate mutual information}}
The conditional mutual information can be used to inductively define a '''multivariate mutual information''' in a set- or [[Information theory and measure theory|measure-theoretic sense]] in the context of '''[[information diagram]]s'''. In this sense we define the multivariate mutual information as follows:
:<math>I(X_1;\cdots;X_{n+1}) = I(X_1;\cdots;X_n) - I(X_1;\cdots;X_n|X_{n+1}),</math>
where
:<math>I(X_1;\cdots;X_n|X_{n+1}) = \mathbb E_{X_{n+1}}\big(I(X_1;\cdots;X_n)|X_{n+1}\big).</math>
This definition is identical to that of [[interaction information]] except for a change in sign in the case of an odd number of random variables. A complication is that this multivariate mutual information (as well as the interaction information) can be positive, negative, or zero, which makes this quantity difficult to interpret intuitively. In fact, for ''n'' random variables, there are <math>2^n-1</math> degrees of freedom for how they might be correlated in an information-theoretic sense, corresponding to each non-empty subset of these variables. These degrees of freedom are bounded by various Shannon- and non-Shannon-type [[inequalities in information theory]].
 
==References==
<references/>
 
[[Category:Information theory]]
[[Category:Entropy and information]]

Revision as of 14:01, 7 September 2012

In probability theory, and in particular, information theory, the conditional mutual information is, in its most basic form, the expected value of the mutual information of two random variables given the value of a third.

Definition

For discrete random variables X, Y, and Z, we define

I(X;Y|Z)=𝔼Z(I(X;Y)|Z)=zZpZ(z)yYxXpX,Y|Z(x,y|z)logpX,Y|Z(x,y|z)pX|Z(x|z)pY|Z(y|z),

where the marginal, joint, and/or conditional probability mass functions are denoted by p with the appropriate subscript. This can be simplified as

I(X;Y|Z)=zZyYxXpX,Y,Z(x,y,z)logpZ(z)pX,Y,Z(x,y,z)pX,Z(x,z)pY,Z(y,z).

Alternatively, we may write[1]

I(X;Y|Z)=H(X,Z)+H(Y,Z)H(X,Y,Z)H(Z)=H(X|Z)H(X|Y,Z)

This can be rewritten to show its relationship to mutual information

I(X;Y|Z)=I(X;Y,Z)I(X;Z)

usually rearranged as the chain rule for mutual information

I(X;Y,Z)=I(X;Z)+I(X;Y|Z)

Another equivalent form of the above is

I(X;Y|Z)=H(Z|X)+H(X)+H(Z|Y)+H(Y)H(Z|X,Y)H(X,Y)H(Z)=I(X;Y)+H(Z|X)+H(Z|Y)H(Z|X,Y)H(Z)

Conditioning on a third random variable may either increase or decrease the mutual information: that is, the difference I(X;Y|Z)I(X;Y), called the interaction information, may be positive, negative, or zero, but it is always true that

I(X;Y|Z)0

for discrete, jointly distributed random variables X, Y, Z. This result has been used as a basic building block for proving other inequalities in information theory, in particular, those known as Shannon-type inequalities.

Like mutual information, conditional mutual information can be expressed as a Kullback-Leibler divergence:

I(X;Y|Z)=DKL[p(X,Y,Z)p(X|Z)p(Y|Z)p(Z)].

Or as an expected value of simpler Kullback-Leibler divergences:

I(X;Y|Z)=zZp(Z=z)DKL[p(X,Y|z)p(X|z)p(Y|z)],
I(X;Y|Z)=yYp(Y=y)DKL[p(X,Z|y)p(X|Z)p(Z|y)].

More general definition

A more general definition of conditional mutual information, applicable to random variables with continuous or other arbitrary distributions, will depend on the concept of regular conditional probability. (See also.[2][3])

Let (Ω,,P) be a probability space, and let the random variables X, Y, and Z each be defined as a Borel-measurable function from Ω to some state space endowed with a topological structure.

Consider the Borel measure (on the σ-algebra generated by the open sets) in the state space of each random variable defined by assigning each Borel set the P-measure of its preimage in . This is called the pushforward measure X*P=P(X1()). The support of a random variable is defined to be the topological support of this measure, i.e. suppX=suppX*P.

Now we can formally define the conditional probability measure given the value of one (or, via the product topology, more) of the random variables. Let M be a measurable subset of Ω, (i.e. M,) and let xsuppX. Then, using the disintegration theorem:

P(M|X=x)=limUxP(M{XU})P({XU})andP(M|X)=MdP(ω|X=X(ω)),

where the limit is taken over the open neighborhoods U of x, as they are allowed to become arbitrarily smaller with respect to set inclusion.

Finally we can define the conditional mutual information via Lebesgue integration:

I(X;Y|Z)=ΩlogdP(ω|X,Z)dP(ω|Y,Z)dP(ω|Z)dP(ω|X,Y,Z)dP(ω),

where the integrand is the logarithm of a Radon–Nikodym derivative involving some of the conditional probability measures we have just defined.

Note on notation

In an expression such as I(A;B|C), A, B, and C need not necessarily be restricted to representing individual random variables, but could also represent the joint distribution of any collection of random variables defined on the same probability space. As is common in probability theory, we may use the comma to denote such a joint distribution, e.g. I(A0,A1;B1,B2,B3|C0,C1). Hence the use of the semicolon (or occasionally a colon or even a wedge ) to separate the principal arguments of the mutual information symbol. (No such distinction is necessary in the symbol for joint entropy, since the joint entropy of any number of random variables is the same as the entropy of their joint distribution.)

Multivariate mutual information

Mining Engineer (Excluding Oil ) Truman from Alma, loves to spend time knotting, largest property developers in singapore developers in singapore and stamp collecting. Recently had a family visit to Urnes Stave Church. The conditional mutual information can be used to inductively define a multivariate mutual information in a set- or measure-theoretic sense in the context of information diagrams. In this sense we define the multivariate mutual information as follows:

I(X1;;Xn+1)=I(X1;;Xn)I(X1;;Xn|Xn+1),

where

I(X1;;Xn|Xn+1)=𝔼Xn+1(I(X1;;Xn)|Xn+1).

This definition is identical to that of interaction information except for a change in sign in the case of an odd number of random variables. A complication is that this multivariate mutual information (as well as the interaction information) can be positive, negative, or zero, which makes this quantity difficult to interpret intuitively. In fact, for n random variables, there are 2n1 degrees of freedom for how they might be correlated in an information-theoretic sense, corresponding to each non-empty subset of these variables. These degrees of freedom are bounded by various Shannon- and non-Shannon-type inequalities in information theory.

References

  1. K. Makarychev et al. A new class of non-Shannon-type inequalities for entropies. Communications in Information and Systems, Vol. 2, No. 2, pp. 147–166, December 2002 PDF
  2. Regular Conditional Probability on PlanetMath
  3. D. Leao Jr. et al. Regular conditional probability, disintegration of probability and Radon spaces. Proyecciones. Vol. 23, No. 1, pp. 15–29, May 2004, Universidad Católica del Norte, Antofagasta, Chile PDF