Tripling-oriented Doche–Icart–Kohel curve

From formulasearchengine
Jump to navigation Jump to search

The Randomized weighted majority algorithm is an algorithm that belongs to the machine learning theory. It minimize the mistake-bound of the weighted majority algorithm. For example, each day we get a prediction from some number of experts about whether the stock market is up or down.Then we use them to make our prediction. Our goal is to do nearly as well as best of the experts in hindsight.

Motivation

In machine learning, the weighted majority algorithm(WMA) is a meta-learning algorithm which "predicts from expert advice".

The algorithm:

initialize all experts to weight 1.
for each round:
    poll all the experts and predict based on a weighted majority vote of their predictions.
    cut in half the weights of all experts that make a mistake.

Suppose there are n experts and the best experts made m mistakes. the weighted majority algorithm makes at most ) mistakes.

Randomized weighted majority(RWMN)

) not so good if the best expert makes a mistake 20% of the time. can we do better? yes. we'd like to improve dependence on . Instead of predicting based on majority vote, we use the weighted as probabilities - Randomized weighted majority. If is the weight of expert , , then we will follow expert with probability . Our goal is to bound the worst-case expected number of mistakes, assuming that the adversary(the world) has to select one of the answers as correct before we make our coin toss. Why is this better in the worst case? Idea: worst case for determinisric algorithm(weighted majority algorithm) was when weights split 50/50.But, now it is not so bad since we also have 50/50 chance of getting it right. Also, to trade-off between dependence on and , we will generalize to multiply by , instead of necessarily .

Analysis

At the -th round, define to be the fraction of weight on the wrong answers. so, is the probability we make a mistake on the -th round. Let denote the total number of mistakes we made so far. Furthermore, we define , using the fact that expectation is additive. On the -th round, becomes . Reason: on fraction, we are multiplying by . So,
Let's say that is the number of mistakes of the best expert so far. We can use the inequality . Now we solve. First, take the natural log of both sides. We get: , Simplify:
, So,
.

Now, use , and the result is:

Let's see if we made any progress:

If , we get, ,
if , we get, .
so we can see we made progress. Roughly, of the form .

Uses of Randomized weighted Majority(RWMN)

Can use to combine multiple algorithms to do nearly as well as best in hindsight.

can apply Randomized weighted majority algorithm in situations where experts are making choices that cannot be combined (or can't be combined easily).For instance, repeated game-playing or online shortest path problem.In the online shortest path problem, each expert is telling you a different way to drive to work. You pick one using Randomized weighted majority algorithm. Later you find out how well you would have done, and penalize appropriately. To do this right, we want to generalize from just "losS" of 0 to 1 to losses in [0,1]. Goal of having expected loss be not too much worse than loss of best expert.We generalize by penalize , meaning having two examples of loss gives same weight as one example of loss 1 and one example of loss 0 (Analysis still oes through).

Extensions

- "Bandit" problem
- Efficient algorithm for some cases with many experts.
- Sleeping experts/"specialists" setting.

See also

Further reading