Few-Labeled Visual Recognition For Self-Driving Using Multi-View Visual-Semantic Representation

2020 
Abstract For a car self-driving system, it is vital to accurately recognize the objects on the road. This is often achieved by analyzing the images captured by various cameras. However, in real applications, we often have limited number of labeled images. Besides, it is very expensive to manually annotate objects in images. It is hard to learn reliable classifiers when only few-labeled images are available. To alleviate these problems, multi-view information is used. However, only using visual information is not enough for reliable classification. Besides, we often have limited images with labels. To cope with these problems, in this paper, we propose a novel multi-view visual-semantic representation method for few-labeled visual recognition ( MV 2 S ). We make use of the state-of-the-art deep convolutional neural networks by viewing them as different views to extract semantic representations of images. This is achieved by using the learned deep convolutional neural networks to make predictions of the semantics of images. Both the visual and semantic representations of images are then used to predict the categories of images by combining the predictions of multi-views with visual and semantic consistency constraints. Experiments on four public available datasets prove the effectiveness of the proposed MV 2 S method.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    66
    References
    1
    Citations
    NaN
    KQI
    []