Margin-infused relaxed algorithm: Difference between revisions
Magioladitis (talk | contribs) m →External links: clean up using AWB (8279) |
m correct cited name |
||
Line 3: | Line 3: | ||
A two-class version called '''binary MIRA'''<ref name="crammer"/> simplifies the algorithm by not requiring the solution of a [[quadratic programming]] problem (see below). When used in an [[one-vs.-all]] configuration, binary MIRA can be extended to a multiclass learner that approximates full MIRA, but may be faster to train. |
A two-class version called '''binary MIRA'''<ref name="crammer"/> simplifies the algorithm by not requiring the solution of a [[quadratic programming]] problem (see below). When used in an [[one-vs.-all]] configuration, binary MIRA can be extended to a multiclass learner that approximates full MIRA, but may be faster to train. |
||
The flow of the algorithm<ref> |
The flow of the algorithm<ref>Watanabe, T. et al (2007): ''Online Large Margin Training for Statistical Machine Translation''. In: Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational |
||
Natural Language Learning, 764–773.</ref><ref>Bohnet, B. (2009): ''Efficient Parsing of Syntactic and Semantic Dependency Structures''. Proceedings of Conference on Natural Language Learning (CoNLL), Boulder, 67-72.</ref> looks as follows: |
Natural Language Learning, 764–773.</ref><ref>Bohnet, B. (2009): ''Efficient Parsing of Syntactic and Semantic Dependency Structures''. Proceedings of Conference on Natural Language Learning (CoNLL), Boulder, 67-72.</ref> looks as follows: |
||
Revision as of 11:23, 9 September 2012
Margin Infused Relaxed Algorithm (MIRA)[1] is a machine learning algorithm, an online algorithm for multiclass classification problems. It is designed to learn a set of parameters (vector or matrix) by processing all the given training examples one-by-one and updating the parameters according to each training example, so that the current training example is classified correctly with a margin against incorrect classifications at least as large as their loss.[2] The change of the parameters is kept as small as possible.
A two-class version called binary MIRA[1] simplifies the algorithm by not requiring the solution of a quadratic programming problem (see below). When used in an one-vs.-all configuration, binary MIRA can be extended to a multiclass learner that approximates full MIRA, but may be faster to train.
The flow of the algorithm[3][4] looks as follows:
Algorithm MIRA Input: Training examples Output: Set of parameters
← 0, ← 0 for ← 1 to for ← 1 to ← update according to ← end for end for return
- "←" denotes assignment. For instance, "largest ← item" means that the value of largest changes to the value of item.
- "return" terminates the algorithm and outputs the following value.
The update step is then formalized as a quadratic programming[2] problem: Find , so that , i.e. the score of the current correct training must be greater than the score of any other possible by at least the loss (number of errors) of that in comparison to .
References
- ^ a b Crammer, K., Singer, Y. (2003): Ultraconservative Online Algorithms for Multiclass Problems. In: Journal of Machine Learning Research 3, 951-991. http://jmlr.csail.mit.edu/papers/v3/crammer03a.html
- ^ a b McDonald, R., K. Crammer and F.C.N. Pereira (2005): Online Large-Margin Training of Dependency Parsers. In: Proceedings of the 43rd Annual Meeting of the ACL, pp. 91-98. http://aclweb.org/anthology-new/P/P05/P05-1012.pdf
- ^ Watanabe, T. et al (2007): Online Large Margin Training for Statistical Machine Translation. In: Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 764–773.
- ^ Bohnet, B. (2009): Efficient Parsing of Syntactic and Semantic Dependency Structures. Proceedings of Conference on Natural Language Learning (CoNLL), Boulder, 67-72.