Margin-infused relaxed algorithm: Difference between revisions

From testwiki
Jump to navigation Jump to search
imported>Rare Voguer
No edit summary
 
(No difference)

Latest revision as of 15:29, 3 July 2024

Template:Short description Margin-infused relaxed algorithm (MIRA)[1] is a machine learning algorithm, an online algorithm for multiclass classification problems. It is designed to learn a set of parameters (vector or matrix) by processing all the given training examples one-by-one and updating the parameters according to each training example, so that the current training example is classified correctly with a margin against incorrect classifications at least as large as their loss.[2] The change of the parameters is kept as small as possible.

A two-class version called binary MIRA[1] simplifies the algorithm by not requiring the solution of a quadratic programming problem (see below). When used in a one-vs-all configuration, binary MIRA can be extended to a multiclass learner that approximates full MIRA, but may be faster to train.

The flow of the algorithm[3][4] looks as follows:

Template:Algorithm-begin

  Input: Training examples T={xi,yi}
  Output: Set of parameters w
  i ← 0, w(0) ← 0
  for n ← 1 to N
    for t ← 1 to |T|
      w(i+1) ← update w(i) according to {xt,yt}
      ii+1
    end for
  end for
  return j=1N×|T|w(j)N×|T|

Template:Algorithm-end

The update step is then formalized as a quadratic programming[2] problem: Find minw(i+1)w(i), so that score(xt,yt)score(xt,y)L(yt,y) y, i.e. the score of the current correct training y must be greater than the score of any other possible y by at least the loss (number of errors) of that y in comparison to y.

References

Template:Reflist

  1. 1.0 1.1 Template:Cite journal
  2. 2.0 2.1 Template:Cite conference
  3. Watanabe, T. et al (2007): "Online Large Margin Training for Statistical Machine Translation". In: Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 764–773.
  4. Bohnet, B. (2009): Efficient Parsing of Syntactic and Semantic Dependency Structures. Proceedings of Conference on Natural Language Learning (CoNLL), Boulder, 67–72.