By a News Reporter-Staff News Editor at Information Technology Newsweekly -- Current study results on Knowledge Discovery have been published. According to news reporting originating in Davis, California, by VerticalNews journalists, research stated, "Active Learning is a machine learning and data mining technique that selects the most informative samples for labeling and uses them as training data; it is especially useful when there are large amount of unlabeled data and labeling them is expensive. Recently, batch-mode active learning, where a set of samples are selected concurrently for labeling, based on their collective merit, has attracted a lot of attention."
The news reporters obtained a quote from the research from the University of California, "The objective of batch-mode active learning is to select a set of informative samples so that a classifier learned on these samples has good generalization performance on the unlabeled data. Most of the existing batch-mode active learning methodologies try to achieve this by selecting samples based on certain criteria. In this article we propose a novel criterion which achieves good generalization performance of a classifier by specifically selecting a set of query samples that minimize the difference in distribution between the labeled and the unlabeled data, after annotation. We explicitly measure this difference based on all candidate subsets of the unlabeled data and select the best subset. The proposed objective is an NP-hard integer programming optimization problem. We provide two optimization techniques to solve this problem. In the first one, the problem is transformed into a convex quadratic programming problem and in the second method the problem is transformed into a linear programming problem. Our empirical studies using publicly available UCI datasets and two biomedical image databases demonstrate the effectiveness of the proposed approach in comparison with the state-of-the-art batch-mode active learning methods. We also present two extensions of the proposed approach, which incorporate uncertainty of the predicted labels of the unlabeled data and transfer learning in the proposed formulation. In addition, we present a joint optimization framework for performing both transfer and active learning simultaneously unlike the existing approaches of learning in two separate stages, that is, typically, transfer learning followed by active learning. We specifically minimize a common objective of reducing distribution difference between the domain adapted source, the queried and labeled samples and the rest of the unlabeled target domain data. Our empirical studies on two biomedical image databases and on a publicly available 20 Newsgroups dataset show that incorporation of uncertainty information and transfer learning further improves the performance of the proposed active learning based classifier."
According to the news reporters, the research concluded: "Our empirical studies also show that the proposed transfer-active method based on the joint optimization framework performs significantly better than a framework which implements transfer and active learning in two separate stages."
For more information on this research see: Batch Mode Active Sampling Based on Marginal Probability Distribution Matching. ACM Transactions on Knowledge Discovery from Data, 2013;7(3):83-107. ACM Transactions on Knowledge Discovery from Data can be contacted at: Assoc Computing Machinery, 2 Penn Plaza, Ste 701, New York, NY 10121-0701, USA.
Our news correspondents report that additional information may be obtained by contacting R. Chattopadhyay, University of California, Davis, CA 95616, United States. Additional authors for this research include Z. Wang, W. Fan, I. Davidson, S. Panchanathan and J.P. Ye.
Keywords for this news article include: Davis, California, United States, Knowledge Discovery, North and Central America
Our reports deliver fact-based news of research and discoveries from around the world. Copyright 2014, NewsRx LLC