Estimate support of a distribution D
Find set
Visualization: take the input space; draw solid ovals around sampled points; the algorithm will draw a dotted oval around these, which will represent the support of the distribution.
With soft margin kernel hyperplane
Aka One Class SVM or OSVM.
Given
Thence get Lagrangian: \
Set derivatives wrt primal vars
Choosing kernel, tuning parameters
With Gaussian kernel, any data set is seperable as everything is mapped to same quadrant in feature space.
\oprob How to decide width of Gaussian kernel to use? Can you use information about the abnormal class in choosing the kernel?
Comparison with thresholded Kernel Density estimator
If
Comparison with using soft margin hyperspheres
For homogenous kernels,
Connection to binary classification
Hyperplane
Using soft margin hyperspheres
Aka Support vector data description. Here one solves:
After using the Lagrangian, finding the critical points and substituting, this leads to the dual
Using Clustering
Cluster the sample, draw boundaries around the clusters. Eg: Use
Novelty detection
Problem
Aka Outlier detection.
In general, we want to find outliers - unlikely data-points according to the conditional distributions
As One class classification
view as a problem where there are multiple classes, but all training examples are from one class only.
Motivation
Outliers are detected either to focus attention on them or to remove them from consideration.
Using density estimation
Do density estimation; call apparently improbable data points novel.
Using support of the distribution
Find distrubution support, call anything outside the support an outlier.
Ransack
One learns a model
Then, one finds
Finally, one repeats the entire procedure till the set of outliers is stable.
Boundary methods
K nearest neighbors
Estimate local density of
Support vector data description
\tbc
PCA
Simplify the data using PCA. \tbc