The winnow algorithm[1] is a technique from machine learning for learning a linear classifier from labeled examples. It is very similar to the perceptron algorithm. However, the perceptron algorithm uses an additive weight-update scheme, while Winnow uses a multiplicative scheme that allows it to perform much better when many dimensions are irrelevant (hence its name winnow). It is a simple algorithm that scales well to high-dimensional data. During training, Winnow is shown a sequence of positive and negative examples. From these it learns a decision hyperplane that can then be used to label novel examples as positive or negative. The algorithm can also be used in the online learning setting, where the learning and the classification phase are not clearly separated.
The basic algorithm, Winnow1, is as follows. The instance space is
X=\{0,1\}n
wi
i\in\{1,\ldots,n\}
(x1,\ldots,xn)
n | |
\sum | |
i=1 |
wixi>\Theta
Here
\Theta
\Theta=n/2
For each example with which it is presented, the learner applies the following update rule:
xi=1
wi
\forallxi=1,wi=0
xi=1
wi
\forallxi=1,wi=\alphawi
A typical value for is 2.
There are many variations to this basic approach. Winnow2[1] is similar except that in the demotion step the weights are divided by instead of being set to 0. Balanced Winnow maintains two sets of weights, and thus two hyperplanes. This can then be generalized for multi-label classification.
In certain circumstances, it can be shown that the number of mistakes Winnow makes as it learns has an upper bound that is independent of the number of instances with which it is presented. If the Winnow1 algorithm uses
\alpha>1
\Theta\geq1/\alpha
k
f(x1,\ldots,xn)=x
i1 |
\cup … \cup
x | |
ik |
\alphak(log\alpha\Theta+1)+
n | |
\Theta |