On Discriminative Semi-Supervised Classification
نویسندگان
چکیده
The recent years have witnessed a surge of interests in semi-supervised learning methods. A common strategy for these algorithms is to require that the predicted data labels should be sufficiently smooth with respect to the intrinsic data manifold. In this paper, we argue that rather than penalizing the label smoothness, we can directly punish the discriminality of the classification function to achieve a more powerful predictor, and we derive two specific algorithms: SemiSupervised Discriminative Regularization (SSDR) and Semi-parametric Discriminative Semi-supervised Classification (SDSC). Finally many experimental results are presented to show the effectiveness of our method. Introduction In many practical pattern classification and data mining problems, the aquisition of sufficient labeled data is often expensive and/or time consuming. However, in many cases, large numbers of unlabeled data are far easier to obtain. For example, in text classification, one may have an easy access to a large database of documents (e.g. by crawling the web), but only a small part of them are classified by hand. Consequently, semi-supervised learning methods, which aim to learn from both labeled and unlabeled data points, are proposed (Chapelle et al., 2006). Many semi-supervised learning algorithms have been proposed in the last decades, for example, the generative model based methods (Shahshahani & Landgrebe, 1994; Miller & Uyar, 1997; Nigam et al, 2000), co-training (Blum & Mitchell, 1998), transductive SVM (TSVM) (Joachims, 1999), and the approaches based on statistical physics theories (Getz, et al, 2005; Wang et al, 2007). One basic assumption behind semi-supervised learning is the cluster assumption (Chapelle et al., 2006), which states that two points are likely to have the same class label if there is a path connecting them passing through the regions of high density only. Zhou et al. (Zhou et al, 2004) further explored the geometric intuition behind this assumption: (1) nearby points are likely to have the same label; (2) points on the same structure (such as a cluster or a submanifold) are Copyright c © 2008, American Association for Artificial Intelligence (www.aaai.org). All rights reserved. The work is supported by NSFC (Grant No. 60721003, 60675009). likely to have the same label. Therefore, the learned classification function should be sufficiently smooth with to respect the intrinsic data manifold. Belkin et al. (Belkin et al., 2006) further proposed an elegant framework for semi-supervised learning, called manifold regularization, which learns a specific classification function by minimizing a cost composed of two terms, one is the structure loss and the other is a smoothness measure estimated from both labeled and unlabeled data. In such a way, the learned function would be sufficiently smooth while simultaneously hold a good generalization ability. However, as the final goal of classification is to discriminate the data points from different classes, it is reasonable that the predicted labels vary smoothly with respect to the intrinsic geometric structure within the same class, but there should definitely be some discontinuities on the decision boundaries, i.e., on the boundaries where the different classes are connected. Moreover, it would be more encouraged if the predicted labels of the data in different classes are more distinct. Based on the above considerations, in this paper, we propose a novel strategy for semi-supervised classification, which does not require the classification function to be smooth, but to be discriminative. We first construct an unsupervised discriminative kernel based on discriminant analysis (Fukunaga, 1990), and then use it to derive two specific algorithms, Semi-Supervised Discriminative Regularization (SSDR) and Semi-parametric Discriminative Semisupervised Classification (SDSC) to realize our strategy. Finally the experimental results on several benchmark data sets are presented to show the effectiveness of our methods. The rest of this paper is organized as follows. In section 2 we will derive an unsupervised discriminative kernel. The main procedure of the SSDR and SDSC algorithms will be introduced in detail in section 3 and section 4. The experimental results will be presented in section 5, followed by the conclusions in section 6. A Discriminative Kernel Given a set of data objectsX = {x1,x2, · · · ,xn}with xi ∈ R , the goal of discriminant analysis (Fukunaga, 1990) is to learn a projection matrix P ∈ R with d ¿ D, such that after projection the data set have a high within class similarity and between-class dissimilarity. Without the loss of Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (2008)
منابع مشابه
Discriminative Similarity for Clustering and Semi-Supervised Learning
Similarity-based clustering and semi-supervised learning methods separate the data into clusters or classes according to the pairwise similarity between the data, and the pairwise similarity is crucial for their performance. In this paper, we propose a novel discriminative similarity learning framework which learns discriminative similarity for either data clustering or semi-supervised learning...
متن کاملA Hybrid Generative/Discriminative Approach to Semi-Supervised Classifier Design
Semi-supervised classifier design that simultaneously utilizes both labeled and unlabeled samples is a major research issue in machine learning. Existing semisupervised learning methods belong to either generative or discriminative approaches. This paper focuses on probabilistic semi-supervised classifier design and presents a hybrid approach to take advantage of the generative and discriminati...
متن کاملA Comparison of Discriminative EM-Based Semi-Supervised Learning algorithms on Agreement/Disagreement classification
Recently, semi-supervised learning has been an active research topic in the natural language processing community, to save effort in hand-labeling for data-driven learning and to exploit a large amount of readily available unlabeled text. In this paper, we apply EM-based semi-supervised learning algorithms such as traditional EM, co-EM, and cross validation EM to the task of agreement/disagreem...
متن کاملSemi-supervised learning for improved expression of uncertainty in discriminative classifiers
Seeking classifier models that are not overconfident and that better represent the inherent uncertainty over a set of choices, we extend an objective for semi-supervised learning for neural networks to two models from the ratio semi-definite classifier (RSC) family. We show that the RSC family of classifiers produces smoother transitions between classes on a vowel classification task, and that ...
متن کاملSemi-supervised Clustering with Pairwise Constraints: A Discriminative Approach
We consider the semi-supervised clustering problem where we know (with varying degree of certainty) that some sample pairs are (or are not) in the same class. Unlike previous efforts in adapting clustering algorithms to incorporate those pairwise relations, our work is based on a discriminative model. We generalize the standard Gaussian process classifier (GPC) to express our classification pre...
متن کاملSemi Supervised Logistic Regression
Semi-supervised learning has recently emerged as a new paradigm in the machine learning community. It aims at exploiting simultaneously labeled and unlabeled data for classification. We introduce here a new semi-supervised algorithm. Its originality is that it relies on a discriminative approach to semisupervised learning rather than a generative approach, as it is usually the case. We present ...
متن کامل