Paper abstractWatch, Listen & Learn: Co-training on Captioned Images and VideosSonal Gupta - University of Texas at Austin, USAJoohyun Kim - University of Texas at Austin, USA Kristen Grauman - University of Texas at Austin, USA Raymond Mooney - University of Texas at Austin, USA Session: Semi Supervised Learning Springer Link: http://dx.doi.org/10.1007/978-3-540-87479-9_48 Recognizing visual scenes and activities is challenging: often visual cues alone are ambiguous, and it is expensive to obtain manually labeled examples from which to learn. To cope with these constraints, we propose to leverage the text that often accompanies visual data to learn robust models of scenes and actions from partially labeled collections. Our approach uses co-training, a semi-supervised learning method that accommodates multi-modal views of data. To classify images, our method learns from captioned images of natural scenes; and to recognize human actions, it learns from videos of athletic events with commentary. We show that by exploiting both multi-modal representations and unlabeled data our approach learns more accurate image and video classifiers than standard baseline algorithms. |