Joint Segmentation and Recognition of Categorized Objects From Noisy Web Image Collection

Le Wang, Gang Hua, Jianru Xue, Zhanning Gao, Nanning Zheng
2014 IEEE Transactions on Image Processing  
The segmentation of categorized objects addresses the problem of joint segmentation of a single category of object across a collection of images, where categorized objects are referred to objects in the same category. Most existing methods of segmentation of categorized objects made the assumption that all images in the given image collection contain the target object. In other words, the given image collection is noise free. Therefore, they may not work well when there are some noisy images,
more » ... ich are not in the same category, such as those image collections gathered by a text query from modern image search engines. To overcome this limitation, we propose a method for automatic segmentation and recognition of categorized objects from noisy Web image collections. This is achieved by cotraining an automatic object segmentation algorithm that operates directly on a collection of images, and an object category recognition algorithm that identifies which images contain the target object. The object segmentation algorithm is trained on a subset of images from the given image collection, which are recognized to contain the target object with high confidence, whereas training the object category recognition model is guided by the intermediate segmentation results obtained from the object segmentation algorithm. This way, our cotraining algorithm automatically identifies the set of true positives in the noisy Web image collection, and simultaneously extracts the target objects from all the identified images. Extensive experiments validated the efficacy of our proposed approach on four data sets: 1) the Weizmann horse data set; 2) the MSRC object category data set; 3) the iCoseg data set; and 4) a new 30-categories data set, including 15 634 Web images with both hand-annotated category labels and ground truth segmentation labels. It is shown that our method compares favorably with the state-of-the-art, and has the ability to deal with noisy image collections.
doi:10.1109/tip.2014.2339196 pmid:25051553 fatcat:47pqanh4zbfcvfxdrcrwre6sjm