Pure spinor: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>Headbomb
 
Line 1: Line 1:
Alyson is what my husband loves to contact me but I don't like when people use my full title. The favorite hobby for him and his children is style and he'll be beginning something else alongside with it. My wife and I live in Mississippi and I love each day living here. She works as a journey agent but soon she'll be  [http://conniecolin.com/xe/community/24580 psychic readers] on her personal.<br><br>Here is my web blog :: tarot card [http://black7.mireene.com/aqw/5741 best psychic readings] ([http://brazil.amor-amore.com/irboothe http://brazil.amor-amore.com/])
In [[statistical classification]], the '''Fisher kernel''', named in honour of Sir [[Ronald Fisher]], is a function that [[Similarity measure|measures the similarity]] of two objects on the basis of sets of measurements for each object and a statistical model. In a classification procedure, the class for a new object (whose real class is unknown) can be estimated by minimising, across classes, an average of the Fisher kernel distance from the new object to each known member of the given class.
 
The Fisher kernel was introduced in 1998.<ref>
Tommi Jaakkola and David Haussler (1998), Exploiting Generative Models in Discriminative Classifiers. In ''Advances in Neural Information Processing Systems 11'', pages 487&ndash;493. MIT Press. ISBN 978-0-262-11245-1 [http://people.csail.mit.edu/tommi/papers/gendisc.ps PS], [http://citeseer.ist.psu.edu/jaakkola98exploiting.html Citeseer]</ref> It combines the advantages of [[Generative model|generative statistical models]] (like the [[hidden Markov model]]) and those of [[Statistical classification|discriminative methods]] (like [[support vector machine]]s):
* generative models can process data of variable length (adding or removing data is well-supported)
* discriminative methods can have flexible criteria and yield better results.
 
== Derivation ==
=== Fisher score ===
The Fisher kernel makes use of the '''Fisher [[Score (statistics)|score]]''', defined as
 
: <math>
U_X = \nabla_{\theta} \log P(X|\theta)
</math>
 
with ''θ'' being a set (vector) of parameters. The function taking ''θ'' to log&nbsp;P(''X''|''θ'') is the [[log-likelihood]] of the probabilistic model.
 
=== Fisher kernel ===
The '''Fisher kernel''' is defined as
 
: <math>
K(X_i, X_j) = U_{X_i}^T I^{-1} U_{X_j}
</math>
 
with ''I'' the [[Fisher information]] matrix.
 
== Applications ==
=== Information retrieval ===
The Fisher kernel is the kernel for a generative probabilistic model. As such, it constitutes a bridge between generative and probabilistic models of documents.<ref>Cyril Goutte, Eric Gaussier, Nicola Cancedda, Hervé Dejean (2004))[http://www.xrce.xerox.com/Research-Development/Publications/2003-0794 "Generative vs Discriminative Approaches to Entity Recognition from Label-Deficient Data"] ''JADT 2004, 7èmes journées internationales analyse statistique des données textuelles'', Louvain-la-Neuve, Belgium, 10-12 mars 2004</ref> Fisher kernels exist for numerous models, notably [[tf–idf]],<ref>{{cite conference |author=Charles Elkan |title=Deriving TF-IDF as a fisher kernel |year=2005 |conference=SPIRE |url=http://lvk.cs.msu.su/~bruzz/articles/not_processed/spire05.pdf}}</ref> [[Naive Bayes]] and [[probabilistic latent semantic analysis]].
 
=== Image classification and retrieval ===
The Fisher kernel can also be applied to image representation for classification or retrieval problems. Currently, the most popular [[Bag of words model in computer vision|bag-of-visual-words]] representation suffers from sparsity and high dimensionality. The Fisher kernel can result in a compact and dense representation, which is more desirable for image classification<ref>Florent Perronnin and Christopher Dance (2007), “Fisher Kernels on Visual Vocabularies for Image Categorization”</ref> and retrieval<ref>Herve Jegou et al. (2010), “Aggregating local descriptors into a compact image representation”</ref> problems.
 
== See also ==
* [[Fisher information metric]]
 
== Notes and references ==
<references/>
 
* Nello Cristianini and John Shawe-Taylor. ''An Introduction to Support Vector Machines and other kernel-based learning methods''. Cambridge University Press, 2000. ISBN 0-521-78019-5 ''([http://www.support-vector.net] SVM Book)''
 
[[Category:Kernel methods for machine learning]]

Revision as of 21:13, 9 December 2013

In statistical classification, the Fisher kernel, named in honour of Sir Ronald Fisher, is a function that measures the similarity of two objects on the basis of sets of measurements for each object and a statistical model. In a classification procedure, the class for a new object (whose real class is unknown) can be estimated by minimising, across classes, an average of the Fisher kernel distance from the new object to each known member of the given class.

The Fisher kernel was introduced in 1998.[1] It combines the advantages of generative statistical models (like the hidden Markov model) and those of discriminative methods (like support vector machines):

  • generative models can process data of variable length (adding or removing data is well-supported)
  • discriminative methods can have flexible criteria and yield better results.

Derivation

Fisher score

The Fisher kernel makes use of the Fisher score, defined as

with θ being a set (vector) of parameters. The function taking θ to log P(X|θ) is the log-likelihood of the probabilistic model.

Fisher kernel

The Fisher kernel is defined as

with I the Fisher information matrix.

Applications

Information retrieval

The Fisher kernel is the kernel for a generative probabilistic model. As such, it constitutes a bridge between generative and probabilistic models of documents.[2] Fisher kernels exist for numerous models, notably tf–idf,[3] Naive Bayes and probabilistic latent semantic analysis.

Image classification and retrieval

The Fisher kernel can also be applied to image representation for classification or retrieval problems. Currently, the most popular bag-of-visual-words representation suffers from sparsity and high dimensionality. The Fisher kernel can result in a compact and dense representation, which is more desirable for image classification[4] and retrieval[5] problems.

See also

Notes and references

  1. Tommi Jaakkola and David Haussler (1998), Exploiting Generative Models in Discriminative Classifiers. In Advances in Neural Information Processing Systems 11, pages 487–493. MIT Press. ISBN 978-0-262-11245-1 PS, Citeseer
  2. Cyril Goutte, Eric Gaussier, Nicola Cancedda, Hervé Dejean (2004))"Generative vs Discriminative Approaches to Entity Recognition from Label-Deficient Data" JADT 2004, 7èmes journées internationales analyse statistique des données textuelles, Louvain-la-Neuve, Belgium, 10-12 mars 2004
  3. 55 years old Systems Administrator Antony from Clarence Creek, really loves learning, PC Software and aerobics. Likes to travel and was inspired after making a journey to Historic Ensemble of the Potala Palace.

    You can view that web-site... ccleaner free download
  4. Florent Perronnin and Christopher Dance (2007), “Fisher Kernels on Visual Vocabularies for Image Categorization”
  5. Herve Jegou et al. (2010), “Aggregating local descriptors into a compact image representation”
  • Nello Cristianini and John Shawe-Taylor. An Introduction to Support Vector Machines and other kernel-based learning methods. Cambridge University Press, 2000. ISBN 0-521-78019-5 ([1] SVM Book)