Mistake bounded models

Mistake bound (MB) learning model

Problem

L learns C: L given sample point x, returns h(x), told if it is correct; this is repeated; has mistake bound m (over any sequence of examples).

Adversarial nature and randomization

The mistake bounded model is adversarial in nature - we are concerned with the number of mistakes made in the worst case. So, randomization helps : the adversary decides on the input before the coin is tossed, so not knowing the algorithm’s output, its attempts to cause mistakes are less successful.

Traits of MB learners

General traits

Any L with finite mb can be written as sober algorithm L which makes its hypothesis consistant with all examples seen so far: Otherwise, can feed inconsistant hypothesis repeatedly, and L would exceed mb.

Careful algorithm Lc: updates hypothesis only if it makes mistake. L without careless updates is Lc, which has same mistake bound: Else, if Lc makes m+1 mistakes, L would exceed mistake bound m on that sequence.

Efficient learnability in MB model

Show runtime per trial = poly(n, size(c)), show polynomial mistake bounds. Reduce to efficiently learnable problem.

Learnability in EQ only model

MB oracle can be simulated using an EQ oracle.

Learnability in PAC model

Any C which is efficiently learnable in the MB model is PAC learnable. We use this PAC algorithm: Take sober algorithm, get hypothesis consistant with large enough sample S drawn from the given distribution D.

Or translate to EQ algorithm, then convert to PAC algorithm.

Lower bound

mb=Ω(VCD(Cr,n)): Else you’d be able to learn in the PAC model with an impossibly low sample complexity.

Halving algorithm

For |C| finite: At max O(log|C|) mistakes needed, ignoring efficiency: L replies to x with maj(h1(x),) : falsifies atleast 1/2 the concepts with every mistake.

mb of Halving algorithm is a lower bound: no algorithm can make better use of sample \chk. mb=Ω(|C|) not lower bound: Take k point functions; halving algorithm has mb = 1, not log k.

\oprob Compare halving algorithm running time with VCD.

Make Mistake bounded learning algorithm for C using H

Decide initial h (maybe null, universal); decide update to hypothesis corresponding to example provided by the sampling oracle: Eg: Learn Disjunctions with mistake bound 2n.

Find mistake bound

Enumerate cases where algorithm makes mistake; find max num of improvements to h possible in each case.

Start with some money; show that you never get to 0.

\oprob In MB model: Does ae learnability imply strong ae learnability?

Mistake bounds (mb) for some R

Disjunctions (so Conjunctions)

n vars, k terms.

Simplified Winnow

Mb = O(k log n), runtime O(n).

Algorithm: Initial h is the sign of the halfspace f=xin of weight W=0; if mistake on +ve x, double weights of all Wt(xi)=1 upto max n; if mistake on -ve x, xi=1, set Wt(xi)=0.

Analysis: On +ve x: From f defn, xi=1Wt(xi)n, Max wt added: n, so max mistakes klogn. On -ve x: From f defn, xi=1Wt(xi)n, Min wt removed: n, so mb = klogn.

Decision lists

Length k. 1-Decn lists: Have bag of candidates for 1st position (bag 1); Demote to bag 2 if guess is wrong; etc.. ; mb=O(nk)=O(n2). \oprob: Get closer to Ω(|C|)=Ω(klogn).

So, t-Decision lists efficiently learnable, by \textbf{Feature expansion}; mb = O(ntk)=nO(t). As |C|=ntk,Ω(tklogn)=O(tntlogn) needed.

Decision trees

Rank of dec tree is rlogn, so dec tree reduced to logn+1 dec list learnt with mb= nO(r)=nO(logn). Information theoretically O(n(logn)) enough. \why \oprob: Can we get to O(nk)?. Similarly l augmented dec tree of rank r has mb=nO(l+r).

Polynomial size DNF f

Reduce to l-augmented dec tree of rank 2nllogs+1;\ taking l=nlogs, f reduced to O(nlogs) decn list, so mb = nO(nlogs).

mb=O(nO(n1/3logs)): reduce to O(n1/3logs) degree PTF.

Halfspace

Given sample set S={(xi,yi)}. The target classifier has the form: aixia00i:yi=1. We want to learn some possible ai,a0R. This can be solved with linear programming; max mistakes: O(nc).

See statistics survey for the winnow algorithm, the perceptron algorithm, SVM’s etc.

Learn parity

Use the GF(2) representation of assignments and parity functions. See mq only algorithm to learn parity.

Could be important in learning Fourier coefficients.

\oprob Parity: Show learnability in IMB model.

Halfspaces: Sample net algorithm for Uniform distr

Take S labeled points; given random point p, choose points with positive inner products; return label of the majority.

Halfspaces: Averaging algorithm

Problem

With U(μ,σ). Target hyperplane c through origin; unit vector uc defines c. Draw S points uniformly from unit sphere Sn1’s surface.

Algorithm

Reflect xi with c(xi)=1; get all +ve S. uh=avgxiUSxi.

Proof of goodness

Angle between u, uh be θ: Pr(sgn(u,x)sgn(uh,x))=θπ. Let \(u’{h}\) be component of \(u{h} \perp u\). θ=tan1uhu,uh.

Show uh,u large; \ so see uh,u=m1uh,xi; so E[uh,u]=E[uh,xi]. But, for u fixed and x uniform from unit sphere, \ Pr(au,xb)=nab(1z2n3dz) \why. So, E[uh,xi]=2n01(1z2)n32zdz2n01n(1z2)n32zdzc. Also, can see E[uh,u] big whp. \why

Show \(u’{h}\) small: Wlog, take u=(1,0..,0), let \(u{h}’ = v\). v1=0; get other vi by choosing points uniformly at random from Sn2: get each viN(0,n1). So, each vi=O(tn) whp. uhm0.5. \why

So, θ=nmπ. \chk. For m=nϵ2, error = θπϵπ.

\oprob Prove that of halfspaces using averaging algorithm works for arbit distr.

PTF of degree d

mb=O(nd), Time: O(ndc)=O(nO(d)): reduce to Halfspace.

Of a panel of k experts

The problem

You have a panel of experts (ei), who on input x return their verdict (ei(x)) on whether c(x)=1 or c(x)=1.

Best expert for input sequence

It is possible that no ei(x) works perfectly forallx. So, for every ei, there is always an input sequence, for which ei(x) is always wrong. However, for a given input sequence, the ‘best expert’ can easily discerned from hindsight.

Performance goal

If you knew who the least erroneous expert was, you would use the verdict of that expert alone; but you do not know this. You want an algorithm which combines the verdicts of all these experts in such a way that you do not perform much worse than the best expert.

Weighted majority

Classifier returns sgn(iwiei).

Learning algorithm: init: wi=1; mb(best expert) = mmin. When wrong, halve wi.

Analysis: When the algorithm is wrong, 12 of total weight W is on wrong experts and 14W is lost due to the corrective update. So, after m mistakes, W(34)mk. So, considering the weight of the best expert after m mistakes, (12)mminW(0.75)mk; so mO(mmin+logk).

Matter of focusing wt on the best expert: see also external regret minimization analysis in Game Theory ref.

Randomized weighted majority

In this version of the algorithm, the classifier returns +1 with probability i:ei(x)=1wiiwi. This is the same as returning the vote of the expert ei with probability wi.

The expected number of mistakes after a certain number of rounds (rather than a certain number of mistakes) can then be analysed using a similar analysis. It turns out to be better. For intuition as to why this is the case, see note on the adversarial nature of the mistake bounded learning model.

Comparison with halfspace algorithms

By vieweing input bits as experts, learning a halfspace wTx+w0 used to classify x can be cast as one of learning weights to be assigned to a panel of experts.

For comparison with some related half-space learning algorithms, like the winnow and the perceptron learning algorithm, see sections about them in the statistics survey.

Intersection F of k halfspaces

As NSaka. \why

Infinite attribute MB learning model (IMB)

Problem and notation

Infinite literals {xi} out there. A sample point is specified by a list of attributes present in the sample.

r is the number of variables c actually depends on. L: MB model algorithm for learning C. n, when not used in the MB sense: size of the largest example seen. L : algorithm to learn in infinite attribute model. p(r,|c|)h(n) = mistake bound of L; p(r,|c|)h(n) = mistake bound of L. Ni : set of literals used by L at step i.

We can assume that L knows r and p: else, r and p can be doubled or squared till L stops making more than p(r,|c|)h(n) mistakes.

The key problem

Which attributes are relevant for classification?

Importance

Eg: A rover in mars trying to understand the properties of life there.

ae learning

Same as in MB model, except use n as size of largest example.

Lower bound

Using MB model lower bounds: Ω(VCD(Cr,n)).

Sequential learning algorithm

L uses L and an attribute set N to label the examples.

Algorithm

Init: N empty. When L makes p(r,|c|)h(|Ni|) mistakes, we know that Ni doesn’t have all relevant vars; during mistakes, we would have seen 1 relevant literal not already in Ni. So, we set Ni+1 to include all variables seen during mistakes, along with Ni.

Analysis

After r iterations, L gets all relevant literals. So, mistake bound is O(rp(r,|c|)h(|Nr|))=rp(r,|c|)h(n)p(r,|c|)h(|Nr1|)))

Learning by mapping

Works for pec C. Keep mapping/ projecting variables to a list of variables, N: yields better bound, simpler analysis. We use s=|c| below.

Whenever a mistake is made, any new attribute is immediately added to N. Now, if the MB algorithm made at most p(r,s)h(|N|) mistakes, the new algorithm makes at most p(r,s)h(|N|) mistakes, where |N| solves |N|np(r,s)h(|N|)+n.

So, if pec class ae learnable in MB model, it is also learnable in the IMB model. Also, if pec class strongly ae learnable in MB model, it is also strongly ae learnable in the IMB model.

\oprob Show that learnability in IMB implies ae learnability in MB model.

Results

Using O(rlogn) winnow algorithm from MB, disjunctions \ learnable with mb O(r3logrn) and time per trial O(rnlogrn). So, size s k-CNF and k-DNF learnable using winnow with time per trial O~(nk) and mb O(s3klogsn).

\oprob Learn decision lists in the IMB in time and mistake bound poly(n, r)?

Expensive Sequential halving algorithm

Sequential learning algorithm where L = expensive halving algorithm with mb=log|C(N)|, C(N)={c|c(x)=c(xN}. C(r)=C(S):|S|[0,r]. C(N)(|N|k)C(k) as all c based on max r literals. So, at any step, p(r,|c|)h(n)=rlog(|N||C(r)|). So, p(r,|c|)h(n)=r2log(nr|C(r)|).

MBQ and IMBQ models

MB and IMB models with MQ oracles. Trying to learn pec class. An example of Active Learning. Motivating scenarios: password cracking.

If f(s1)=1 and f(s2)=0, with logn membership queries (mq), you can identify the relevant variable in s1Δs2.

Lower bound: Ω(VCD(Cr,n)). \why

Convert an efficient MBQ algorithm (L) with bound q(n,|c|) into an algorithm L that strongly ae learns the class in MBQ or the IMBQ model with bound 2(r+1)q(n,|c|)+r(logm+1). Pf: N: set of possibly relevant attributes; Init: N=ϕ; keep growing N to include all relevant attributes. Take arbit partial assignment P for V-N; Run L to learn cP. Keep projecting examples (sP/s) and mq to and fro; Respond to mq in the obvious way. If it makes a mistake on some s, check using mq if c(s)c(P/s); if so, find and add the relevant variable therein. So, bound of the resultant algorithm L is roughly r × bound of L: , L is ae.

Predictive power of Consistent (maybe small) h

(a,b) Occam algorithm for C using H

Give h consistant with sample S, size(h)(n.size(c))amb.

So, the size of the h generated is not too big. This property turns out to be important in bounding the size of H, which in turn helps us prove the goodness of the PAC learning algorithm which we craft using the Occamish algorithm.

This is reminiscent of the predictive power of Occam razor used in elucidating the philosophy of science after the rennaissance: make as few assumptions in your theory as possible.

Occam razor: Goodness from consistency

Any efficient Occam algorithm can be used to construct a good PAC algorithm: simply draw m=O(ϵ1(log|Hn,m|+log(δ1))) samples from the distribution D and use the Occam algorithm to learn a hypothesis h consistent with this sample set.

\proof: Using the Chernoff bounds, we can say: The empirical estimate of the goodness of a fixed h on a large sample set S is, with high probability, very close to its actual goodness. But there can be many h which are ϵ close to c, and we want to be able to say that, irrespective of which h the Occam algorithm outputs, it is likely to be good. To do this, we use the union bound from probability theory: Pr( bad h;[cΔh]S=ϕ)|Δϵ(c)|(1ϵ)m|H|(1ϵ)m|H|emϵδ.

In other words, we found a way of saying that S is likely to be an ϵ net around c.

VCD - Occam razor: Extension to unbounded C

Even if |H|=, m=Ω(ϵ1log(δ1)+dϵlog(ϵ1)) examples likely to form ϵ net: use bound on ΠC(m). So, whp:\ ϵ=O(m1log(ΠC(2m))+m1logd1): ϵ decreases with increase in m, decrease in ΠC(2m).

Almost consistent h

Suppose that an Occam-like algorithm produces a h with a small, non zero empirical error on S: ϵS.

Using the Hoeffding’s inequalityuality, for a fixed h, we see that this is likely to be a good estimate of the actual error. Pr(|ϵ2ϵS|ϵ2)2eϵ22m; so any h with ϵS=ϵ2 on m=Ω(ϵ2log1δ) good whp.

As before, the union bound using either the VCD d or using |H| can then be applied to get: Ω(ϵ2(log|H|+log1δ)) and Ω(ϵ2(d+log1δ)).

Occam with Approximate set cover

Take set U of m examples seen so far; |c| := size (num of literals) of smallest c to cover U. Greedily, repeatedly alter c (add literal) to cover most of uncovered part of U at step i (Ui): as c covers Ui, atleast 1 literal in c covers Ui|c|; so Ui+1=|Ui|Ui|c|m(1|c|1)i; so, |h|=O(|c|logmlogn); by Occam razor m=O(|c|log2n+log1dϵ). Eg: learn disjunctions of size k.

Converse to Occam razor

For any pac learnable C, can use Adaboost with L, do boosting by sampling to find small hypothesis consistent with any sample S.