عنوان مقاله فارسی: بازیابی متقابل با ویژگی های بصری CNN: یک خط پایه جدید
عنوان مقاله لاتین: Cross-Modal Retrieval With CNN Visual Features: A New Baseline
نویسندگان: Yunchao Wei; Yao Zhao; Canyi Lu; Shikui Wei; Luoqi Liu; Zhenfeng Zhu; Shuicheng Yan
تعداد صفحات: 11
سال انتشار: 2017
زبان: لاتین
Abstract:
Recently, convolutional neural network (CNN) visual features have demonstrated their powerful ability as a universal representation for various recognition tasks. In this paper, cross-modal retrieval with CNN visual features is implemented with several classic methods. Specifically, off-the-shelf CNN visual features are extracted from the CNN model, which is pretrained on ImageNet with more than one million images from 1000 object categories, as a generic image representation to tackle cross-modal retrieval. To further enhance the representational ability of CNN visual features, based on the pretrained CNN model on ImageNet, a fine-tuning step is performed by using the open source Caffe CNN library for each target data set. Besides, we propose a deep semantic matching method to address the cross-modal retrieval problem with respect to samples which are annotated with one or multiple labels. Extensive experiments on five popular publicly available data sets well demonstrate the superiority of CNN visual features for cross-modal retrieval.
cross-modal retrieval with cnn visual features a new baseline_1622885959_48846_4145_1096.zip4.18 MB |