Co-training is a semi-supervised learning method that is designed to take advantage of the redundancy that is present when the object to be identified has multiple descriptions. Co-training is known to work well when the multiple descriptions are conditional independent given the class of the object. The presence of multiple descriptions of objects in the form of text, images, audio and video in multimedia applications appears to provide redundancy in the form that may be suitable for co-training. In this paper, we investigate the suitability of utilizing text and image data from the Web for co-training. We perform measurements to find indications of conditional independence in the texts and images obtained from the Web. Our measurements suggest that conditional independence is likely to be present in the data. Our experiments, within a relevance feedback framework to test whether a method that exploits the conditional independence outperforms methods that do not, also indicate that better performance can indeed be obtained by designing algorithms that exploit this form of the redundancy when it is present. / Singapore-MIT Alliance (SMA)
Identifer | oai:union.ndltd.org:MIT/oai:dspace.mit.edu:1721.1/7438 |
Date | 01 1900 |
Creators | Zhang, Dell, Lee, Wee Sun |
Source Sets | M.I.T. Theses and Dissertation |
Language | English |
Detected Language | English |
Type | Article |
Format | 148397 bytes, application/pdf |
Relation | Computer Science (CS); |
Page generated in 0.0021 seconds