Co-EM support vector learning
Brefeld, Ulf; Scheffer, Tobias
2004-07-04 00:00:00
Co-EM Support Vector Learning Ulf Brefeld [email protected] Tobias Sche er [email protected] Humboldt-Universit¨t zu Berlin, Department of Computer Science, Unter den Linden 6, 10099 Berlin, Germany a Abstract Multi-view algorithms, such as co-training and co-EM, utilize unlabeled data when the available attributes can be split into independent and compatible subsets. Co-EM outperforms co-training for many problems, but it requires the underlying learner to estimate class probabilities, and to learn from probabilistically labeled data. Therefore, coEM has so far only been studied with naive Bayesian learners. We cast linear classi ers into a probabilistic framework and develop a co-EM version of the Support Vector Machine. We conduct experiments on text classi cation problems and compare the family of semi-supervised support vector algorithms under di erent conditions, including violations of the assumptions underlying multiview learning. For some problems, such as course web page classi cation, we observe the most accurate results reported so far. ers based on independent attribute subsets. These classi ers then provide each other with labels for the unlabeled data. The co-EM algorithm (Nigam & Ghani, 2000) combines multi-view learning with the probabilistic EM approach. This, however, requires the learning algorithm to process probabilistically labeled training data and
http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.pnghttp://www.deepdyve.com/lp/association-for-computing-machinery/co-em-support-vector-learning-Mesrx0Hk6m
Co-EM Support Vector Learning Ulf Brefeld [email protected] Tobias Sche er [email protected] Humboldt-Universit¨t zu Berlin, Department of Computer Science, Unter den Linden 6, 10099 Berlin, Germany a Abstract Multi-view algorithms, such as co-training and co-EM, utilize unlabeled data when the available attributes can be split into independent and compatible subsets. Co-EM outperforms co-training for many problems, but it requires the underlying learner to estimate class probabilities, and to learn from probabilistically labeled data. Therefore, coEM has so far only been studied with naive Bayesian learners. We cast linear classi ers into a probabilistic framework and develop a co-EM version of the Support Vector Machine. We conduct experiments on text classi cation problems and compare the family of semi-supervised support vector algorithms under di erent conditions, including violations of the assumptions underlying multiview learning. For some problems, such as course web page classi cation, we observe the most accurate results reported so far. ers based on independent attribute subsets. These classi ers then provide each other with labels for the unlabeled data. The co-EM algorithm (Nigam & Ghani, 2000) combines multi-view learning with the probabilistic EM approach. This, however, requires the learning algorithm to process probabilistically labeled training data and
To get new article updates from a journal on your personalized homepage, please log in first, or sign up for a DeepDyve account if you don’t already have one.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.