Authors
B Schölkopf,
R Williamson,
AJ Smola,
Publication date
1999
Publisher
Schloss Dagstuhl, Leibniz-Zentrum für Informatik
Total citations
Description
Suppose you are given some dataset drawn from an underlying probability dis-tributionPand you want to estimate a subsetSof input space such that theprobability that a test point drawn fromPlies outside ofSis bounded by somea priori specified 0< ν≤ 1. We propose an algorithm to deal with this problem by trying toestimate afunctionfwhich is positive onSand negative on the complement ofS. Thefunctional form offis given by a kernel expansion in terms of a potentially smallsubset of the training data; it is regularized by controlling the length of the weightvector in an associated feature space. We can prove thatνupper bounds the fraction of outliers (training points outsideofS) and lower bounds the fraction of support vectors. Asymptotically, undersome mild condition onP, both become equalities. The algorithm is a natural extension of the support vector algorithm to the case of unlabelled data.