ACMM: Aligned Cross-Modal Memory for Few-Shot Image and Sentence Matching
Huang, Yan; Wang, Liang
2019-11
会议日期2019.10.26-2019.11.2
会议地点Seoul
关键词Image And Sentence Matching
卷号0
期号0
DOI0
页码0
英文摘要

Image and sentence matching has drawn much attention
recently, but due to the lack of sufficient pairwise data for
training, most previous methods still cannot well associate
those challenging pairs of images and sentences containing
rarely appeared regions and words, i.e., few-shot content.
In this work, we study this challenging scenario as few-shot
image and sentence matching, and accordingly propose an
Aligned Cross-Modal Memory (ACMM) model to memorize
the rarely appeared content. Given a pair of image and sentence,
the model first includes an aligned memory controller
network to produce two sets of semantically-comparable interface
vectors through cross-modal alignment. Then the
interface vectors are used by modality-specific read and update
operations to alternatively interact with shared memory
items. The memory items persistently memorize crossmodal
shared semantic representations, which can be addressed
out to better enhance the representation of few-shot
content. We apply the proposed model to both conventional
and few-shot image and sentence matching tasks, and
demonstrate its effectiveness by achieving the state-of-theart
performance on two benchmark datasets.

源文献作者Kyoung Mu Lee
会议录出版者IEEE
会议录出版地USA
URL标识查看原文
内容类型会议论文
源URL[http://ir.ia.ac.cn/handle/173211/25797]  
专题自动化研究所_智能感知与计算研究中心
作者单位中科院自动化所
推荐引用方式
GB/T 7714
Huang, Yan,Wang, Liang. ACMM: Aligned Cross-Modal Memory for Few-Shot Image and Sentence Matching[C]. 见:. Seoul. 2019.10.26-2019.11.2.
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace