Combining Labeled and Unlabeled Data with CoTraining Avrim
Combining Labeled and Unlabeled Data with Co-Training Avrim Blum Tom Mitchell Madison, 1998
Semi-Supervised Learning • Most of ML assumes labeled training data • But labels can be expensive to maintain • Semi-supervised learning: – Make use of unlabeled data in combination of small amount of labeled data
Some Co-Training Slides from a Tutorial by Avrim Blum
Co-training Many problems have two different sources of info you can use to determine label. E. g. , classifying webpages: can use words on page or words on links pointing to the page. Prof. Avrim Blum My Advisor x - Link info & Text info Prof. Avrim Blum My Advisor x 1 - Link info x 2 - Text info
Co-training Idea: Use small labeled sample to learn initial rules. – E. g. , “my advisor” pointing to a page is a good indicator it is a faculty home page. – E. g. , “I am teaching” on a page is a good indicator it is a faculty home page. my advisor
Co-training Idea: Use small labeled sample to learn initial rules. – E. g. , “my advisor” pointing to a page is a good indicator it is a faculty home page. – E. g. , “I am teaching” on a page is a good indicator it is a faculty home page. Then look for unlabeled examples where one rule is confident and the other is not. Have it label the example for the other. hx 1, x 2 i hx 1, x 2 i Training 2 classifiers, one on each type of info. Using each to help train the other.
Results: webpages 12 labeled examples, 1000 unlabeled
Sample Run
Co-Training Theorems • [BM 98] if x 1, x 2 are independent given the label, and if have alg that is robust to noise, then can learn from an initial “weakly-useful” rule plus unlabeled data. “My advisor” Faculty with advisees Faculty home pages
- Slides: 9