Claim
Assume that all losses are in [0,1], the regret of the Multiplicative Weight Algorithm run for T steps with a parameter 0<ϵ≤1/2 is:
RT≤ϵT+ϵlnn In particular, if T>4lnn and we choose ϵ=Tlnn, we have
RT≤2Tlnn
Outline
Define
Wt=i=1∑nwi(t) to be the sum of the weights at time t, at the end of the execution of the algorithm:
- If WT+1 is small, then the offline optimum is large
- If the loss suffered by the algorithm is large, then WT+1 is small
Therefore
- If the algorithm suffers a large loss, then the offline optimum is also large, so regret is small
Call the offline optimum L∗:
L∗=x∈Δmint=1∑Ti=1∑nxi⋅li(t)=i=1,…,nmint=1∑Tli(t) and we can use Lt to denote the loss incurred by the algorithm at time t
Lt=i=1∑nxi(t)li(t)
Lemmas
Lemma 2
If WT+1 is small, then the offline optimum is large
WT+1≥(1−ϵ)L∗ Proof
Let j be any strategy, then
WT+1=i=1∑nwi(T+1)≥wj(T+1)=t=1∏T(1−ϵ)lj(t)=(1−ϵ)∑t=1Tlj(t) Now let i∗ be an offline optimal strateggy,
L∗=t=1∑Tli∗(t) And observe that for i∗,
WT+1≥(1−ϵ)∑t=1Tli∗(t)=(1−ϵ)L∗ Lemma 3
If the loss suffered by the algorithm is large, then WT+1 is small
WT+1≤n⋅t=1∏T(1−ϵLt) Proof
We know W1=n, so we can prove
WT+1≤Wt⋅(1−ϵLt) Observe that
Wt+1=i=1∑nwi(t+1)=i=1∑nwi(t)⋅(1−ϵ)li(t)≤i=1∑nwi(t)⋅(1−ϵ⋅li(t))∵∀z,ϵ∈[0,1],(1−ϵ)z≤1−ϵz=Wti=1∑nxi(t)⋅(1−ϵ⋅li(t))∵xi(t)=wi(t)/Wt=Wt⋅(i=1∑nxi(t)−i=1∑nϵ⋅xi(t)⋅li(t))=Wt⋅(1−ϵLt)
Proof
LEMMA 2: If WT+1 is small, then the offline optimum is large
WT+1≥(1−ϵ)L∗ LEMMA 3: If the loss suffered by the algorithm is large, then WT+1 is small
WT+1≤n⋅t=1∏T(1−ϵLt) Putting them together
(1−ϵ)L∗L∗ln(1−ϵ)≤n⋅t=1∏T(1−ϵLt)≤lnn+t+1∑Tln(1−ϵLt) Note that because of tarylor series ln(1−z)=∑n=1∞−nzn, we have inequality
∀0≤z≤1/2,−z−z2≤ln(1−z)≤−z Then apply this inequality
L∗⋅(−ϵ−ϵ2)≤lnn−ϵ⋅i=1∑TLt Divide by ϵ and rearrange a bit,
i=1∑TLt−L∗≤ϵL∗+ϵlnn≤ϵT+ϵlnn