Co-training is a machine learning algorithm used when there are only small amounts of labeled data and large amounts of unlabeled data. One of its uses is in text mining for search engines. It was introduced by Avrim Blum and Tom Mitchell in 1998.
Algorithm design
Co-training is a semi-supervised learning technique that requires two views of the data. It assumes that each example is described using two different sets of features that provide complementary information about the instance. Ideally, the two views are conditionally independent (i.e., the two feature sets of each instance are conditionally independent given the class) and each view is sufficient (i.e., the class of an instance can be accurately predicted from each view alone). Co-training first learns a separate classifier for each view using any labeled examples. The most confident predictions of each classifier on the unlabeled data are then used to iteratively construct additional labeled training data.[1]
The original co-training paper described experiments using co-training to classify web pages into "academic course home page" or not; the classifier correctly categorized 95% of 788 web pages with only 12 labeled web pages as examples.[2] The paper has been cited over 1000 times, and received the 10 years Best Paper Award at the 25th International Conference on Machine Learning (ICML 2008), a renowned computer science conference.[3][4]
Krogel and Scheffer showed in 2004 that co-training is only beneficial if the data sets are independent; that is, if one of the classifiers correctly labels a data point that the other classifier previously misclassified. If the classifiers agree on all unlabeled data, i.e. they are dependent, labeling the data does not create new information. In an experiment where dependence of the classifiers was greater than 60%, results worsened.[5]
Uses
Co-training has been used to classify web pages using the text on the page as one view and the anchor text of hyperlinks on other pages that point to the page as the other view. Simply put, the text in a hyperlink on one page can give information about the page it links to.[2] Co-training can work on "unlabeled" text that has not already been classified or tagged, which is typical for the text appearing on web pages and in emails. According to Tom Mitchell, "The features that describe a page are the words on the page and the links that point to that page. The co-training models utilize both classifiers to determine the likelihood that a page will contain data relevant to the search criteria." Text on websites can judge the relevance of link classifiers, hence the term "co-training". Mitchell claims that other search algorithms are 86% accurate, whereas co-training is 96% accurate.[6]
Co-training was used on FlipDog.com, a job search site, and by the U.S. Department of Labor, for a directory of continuing and distance education.[6] It has been used in many other applications, including statistical parsing and visual detection.[7]
References
- ↑ Blum, A., Mitchell, T. Combining labeled and unlabeled data with co-training. COLT: Proceedings of the Workshop on Computational Learning Theory, Morgan Kaufmann, 1998, p. 92-100.
- 1 2 Committee on the Fundamentals of Computer Science: Challenges and Opportunities, National Research Council (2004). "6: Achieving Intelligence". Computer Science: Reflections on the Field, Reflections from the Field. The National Academies Press. ISBN 978-0-309-09301-9.
- ↑ McCallum, Andrew (2008). "Best Papers Awards". ICML Awards. Retrieved 2009-05-03.
- ↑ Shavik, Jude (2008). "10 Year Best Paper: Combining labeled and unlabled data with co-training". ICML Awards. Retrieved 2009-05-03.
- ↑ Krogel, Marc-A; Tobias Scheffer (2004). "Multi-Relational Learning, Text Mining, and Semi-Supervised Learning for Functional Genomics" (PDF). Machine Learning. 57: 61–81. doi:10.1023/B:MACH.0000035472.73496.0c.
- 1 2 Aquino, Stephen (24 April 2001). "Search Engines Ready to Learn". Technology Review. Retrieved 2009-05-03.
- ↑ Xu, Qian; Derek Hao Hu; Hong Xue; Weichuan Yu; Qiang Yang (2009). "Semi-supervised protein subcellular localization". BMC Bioinformatics. 10 (Suppl 1): S47. doi:10.1186/1471-2105-10-S1-S47. ISSN 1471-2105. PMC 2648770. PMID 19208149.
- Notes
- Chakrabarti, Soumen (2002). Mining the Web: Discovering Knowledge from Hypertext Data. Morgan-Kaufmann Publishers. p. 352. ISBN 978-1-55860-754-5.
- Nigam, Kamal; Rayid Ghani (2000). "Analyzing the Effectiveness and Applicability of Co-training". Proceedings of the Ninth International Conference on Information and Knowledge Management. NY, USA: ACM: 86–93. CiteSeerX 10.1.1.37.4669.
- Abney, Steven (2007). Semisupervised Learning for Computational Linguistics. CRC Computer Science & Data Analysis. Chapman & Hall. p. 308. ISBN 978-1-58488-559-7.
- Wang, William Yang; Kapil Thadani; Kathleen McKeown (2011). Identifying Event Descriptions using Co-training with Online News Summaries (PDF). the 5th International Joint Conference on Natural Language Processing (IJCNLP 2011). AFNLP & ACL.