Partial Visual-Tactile Fused Learning for Robotic Object Recognition
Zhang T(张涛)1,2,3; Cong Y(丛杨)1,2; Dong JH(董家华)1,2,3; Hou DD(侯冬冬)1,2,3
刊名IEEE Transactions on Systems, Man, and Cybernetics: Systems
2021
页码1-13
关键词Intelligent robots partial multiview learning (PMVL) visual-tactile fused sensing (VTFS)
ISSN号2168-2216
产权排序1
英文摘要

Currently, visual-tactile fusion learning for robotic object recognition has achieved appealing performance, due to the fact that visual and tactile data can offer complementary information. However: 1) the distinct gap between vision and touch makes it difficult to fully explore the complementary information, which would further lead to performance degradation and 2) most of the existing visual-tactile fused learning methods assume that visual and tactile data are complete, which is often difficult to be satisfied in many real-world applications. In this article, we propose a partial visual-tactile fused (PVTF) framework for robotic object recognition to address these challenges. Specifically, we first employ two modality-specific (MS) encoders to encode partial visual-tactile data into two incomplete subspaces (i.e., visual subspace and tactile subspace). Then, a modality gap mitigated (MGM) network is adopted to discover modality-invariant high-level label information, which is utilized to generate gap loss and further help updating the MS encoders for relatively consistent visual and tactile subspaces generation. In this way, the huge gap between vision and touch is mitigated, which would further contribute to mine the complementary visual-tactile information. Finally, to achieve data completeness and complementary visual-tactile information exploration simultaneously, a cycle subspace leaning technique is proposed to project the incomplete subspaces into a complete subspace by fully exploiting all the obtainable samples, where complete latent representations with maximum complementary information can be learned. A lot of comparative experiments conducted on three visual-tactile datasets validate the advantage of the proposed PVTF framework, by comparing with state-of-the-art baselines.

资助项目National Key Research and Development Program of China[2019YFB1310300] ; National Nature Science Foundation of China[61821005] ; Liaoning Revitalization Talents Program[XLYC1807053]
WOS研究方向Automation & Control Systems ; Computer Science
语种英语
WOS记录号WOS:000732098300001
资助机构National Key Research and Development Program of China under Grant 2019YFB1310300 ; National Nature Science Foundation of China under Grant 61821005 ; Liaoning Revitalization Talents Program under Grant XLYC1807053
内容类型期刊论文
源URL[http://ir.sia.cn/handle/173321/29385]  
专题沈阳自动化研究所_机器人学研究室
通讯作者Cong Y(丛杨)
作者单位1.State Key Laboratory of Robotics, Shenyang Institute of Automation, Shenyang 110016, China
2.Institute for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China
3.University of Chinese Academy of Sciences, Beijing 100049, China
推荐引用方式
GB/T 7714
Zhang T,Cong Y,Dong JH,et al. Partial Visual-Tactile Fused Learning for Robotic Object Recognition[J]. IEEE Transactions on Systems, Man, and Cybernetics: Systems,2021:1-13.
APA Zhang T,Cong Y,Dong JH,&Hou DD.(2021).Partial Visual-Tactile Fused Learning for Robotic Object Recognition.IEEE Transactions on Systems, Man, and Cybernetics: Systems,1-13.
MLA Zhang T,et al."Partial Visual-Tactile Fused Learning for Robotic Object Recognition".IEEE Transactions on Systems, Man, and Cybernetics: Systems (2021):1-13.
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace