Abstract
We implemented versions of the SVM appropriate for one-class classification in the context of information retrieval. The experiments were conducted on the standard Reuters data set. For the SVM implementation we used both a version of Schölkopf et al. and a somewhat different version of one-class SVM based on identifying “outlier” data as representative of the second-class. We report on experiments with different kernels for both of these implementations and with different representations of the data, including binary vectors, tf-idf representation and a modification called “Hadamard” representation. Then we compared it with one-class versions of the algorithms prototype (Rocchio), nearest neighbor, naive Bayes, and finally a natural one-class neural network classification method based on “bottleneck” compression generated filters. The SVM approach as represented by Schölkopf was superior to all the methods except the neural network one, where it was, although occasionally worse, essentially comparable. However, the SVM methods turned out to be quite sensitive to the choice of representation and kernel in ways which are not well understood; therefore, for the time being leaving the neural network approach as the most robust.
Original language | English |
---|---|
Pages (from-to) | 139-154 |
Number of pages | 16 |
Journal | Journal of Machine Learning Research |
Volume | 2 |
State | Published - 2002 |
Bibliographical note
Publisher Copyright:© 2001 Larry M. Manevitz and Malik Yousef.
Keywords
- Compression Neural Network
- Neural Network
- Positive Information
- SVM
- Support Vector Machine
- Text Retrieval
ASJC Scopus subject areas
- Control and Systems Engineering
- Software
- Statistics and Probability
- Artificial Intelligence