Visual-Tactile Fused Graph Learning for Object Clustering
Zhang T(张涛)1,2,3; Cong Y(丛杨)2,3; Sun G(孙干)2,3; Dong JH(董家华)1,2,3
刊名IEEE Transactions on Cybernetics
2021
页码1-15
关键词Clustering graph learning unsupervised learning visual-tactile fused sensing
ISSN号2168-2267
产权排序1
英文摘要

Object clustering has received considerable research attention most recently. However, 1) most existing object clustering methods utilize visual information while ignoring important tactile modality, which would inevitably lead to model performance degradation and 2) simply concatenating visual and tactile information via multiview clustering method can make complementary information to not be fully explored, since there are many differences between vision and touch. To address these issues, we put forward a graph-based visual-tactile fused object clustering framework with two modules: 1) a modality-specific representation learning module MR and 2) a unified affinity graph learning module MU. Specifically, MR focuses on learning modality-specific representations for visual-tactile data, where deep non-negative matrix factorization (NMF) is adopted to extract the hidden information behind each modality. Meanwhile, we employ an autoencoder-like structure to enhance the robustness of the learned representations, and two graphs to improve its compactness. Furthermore, MU highlights how to mitigate the differences between vision and touch, and further maximize the mutual information, which adopts a minimizing disagreement scheme to guide the modality-specific representations toward a unified affinity graph. To achieve ideal clustering performance, a Laplacian rank constraint is imposed to regularize the learned graph with ideal connected components, where noises that caused wrong connections are removed and clustering labels can be obtained directly. Finally, we propose an efficient alternating iterative minimization updating strategy, followed by a theoretical proof to prove framework convergence. Comprehensive experiments on five public datasets demonstrate the superiority of the proposed framework.

资助项目Major Project of the New Generation of Artificial Intelligence[2018AAA0102905] ; National Natural Science Foundation of China[61821005] ; National Natural Science Foundation of China[62003336] ; Liaoning Revitalization Talents Program[XLYC1807053] ; Nature Foundation of Liaoning Province of China[2020KF-11-01]
WOS关键词FUSION ; RECOGNITION
WOS研究方向Automation & Control Systems ; Computer Science
语种英语
WOS记录号WOS:000732294000001
资助机构Major Project of the New Generation of Artificial Intelligence under Grant 2018AAA0102905 ; National Natural Science Foundation of China under Grant 61821005 and Grant 62003336 ; Liaoning Revitalization Talents Program under Grant XLYC1807053 ; Nature Foundation of Liaoning Province of China under Grant 2020-KF-11-01
内容类型期刊论文
源URL[http://ir.sia.cn/handle/173321/29420]  
专题沈阳自动化研究所_机器人学研究室
通讯作者Cong Y(丛杨)
作者单位1.University of Chinese Academy of Sciences, Beijing 100049, China
2.State Key Laboratory of Robotics, Shenyang Institute of Automation, Shenyang 110016, China
3.Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China
推荐引用方式
GB/T 7714
Zhang T,Cong Y,Sun G,et al. Visual-Tactile Fused Graph Learning for Object Clustering[J]. IEEE Transactions on Cybernetics,2021:1-15.
APA Zhang T,Cong Y,Sun G,&Dong JH.(2021).Visual-Tactile Fused Graph Learning for Object Clustering.IEEE Transactions on Cybernetics,1-15.
MLA Zhang T,et al."Visual-Tactile Fused Graph Learning for Object Clustering".IEEE Transactions on Cybernetics (2021):1-15.
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace