International Journal of Control, Automation and Systems 2022; 20(12): 3996-4004
Published online November 3, 2022
https://doi.org/10.1007/s12555-021-0929-8
© The International Journal of Control, Automation, and Systems
This paper presents a novel grasp generative residual attention network (RANET) for generating antipodal robotic grasp from multi-modal images with the pixel-wise method. To strengthen the generalization ability of unknown objects, this paper proposed a new structure that differs from the previous grasp generative network in that it additionally integrates a coordinate attention mechanism and a symmetrical skip connection, respectively. Using the coordinate attention module to emphasize meaningful information of the feature map and the symmetrical skip connection to remain more fine-grained details of feature. Moreover, a multi atrous convolution module is included in the structure to capture more high-level information, while a hypercolumn feature fusion method is incorporated for getting the best from the complementation of different layers’ features. Through evaluation on public datasets, the result demonstrates that we achieve 98.9% accuracy on the Cornell dataset which is the state-of-the-art performance with real-time speed(∼ 17 ms), meanwhile, we represent a 93.9% accuracy performance on the Jacquard dataset.
Keywords Convolutional neural networks, deep learning, grasping detection, vision.
International Journal of Control, Automation and Systems 2022; 20(12): 3996-4004
Published online December 1, 2022 https://doi.org/10.1007/s12555-021-0929-8
Copyright © The International Journal of Control, Automation, and Systems.
Qian-Qian Hong, Liang Yang*, and Bi Zeng
University of Electronic Science and Technology of China Zhongshan Institute
This paper presents a novel grasp generative residual attention network (RANET) for generating antipodal robotic grasp from multi-modal images with the pixel-wise method. To strengthen the generalization ability of unknown objects, this paper proposed a new structure that differs from the previous grasp generative network in that it additionally integrates a coordinate attention mechanism and a symmetrical skip connection, respectively. Using the coordinate attention module to emphasize meaningful information of the feature map and the symmetrical skip connection to remain more fine-grained details of feature. Moreover, a multi atrous convolution module is included in the structure to capture more high-level information, while a hypercolumn feature fusion method is incorporated for getting the best from the complementation of different layers’ features. Through evaluation on public datasets, the result demonstrates that we achieve 98.9% accuracy on the Cornell dataset which is the state-of-the-art performance with real-time speed(∼ 17 ms), meanwhile, we represent a 93.9% accuracy performance on the Jacquard dataset.
Keywords: Convolutional neural networks, deep learning, grasping detection, vision.
Vol. 23, No. 1, pp. 1~88
Jialong Xie, Botao Zhang*, Qiang Lu, and Oleg Borisov
International Journal of Control, Automation, and Systems 2024; 22(1): 252-264Dong-Han Lee, Kyung-Soo Kwak, and Soo-Chul Lim*
International Journal of Control, Automation, and Systems 2023; 21(12): 4032-4040Jeonghoon Kwak, Kyon-Mo Yang, Ye Jun Lee, Min-Gyu Kim, and Kap-Ho Seo*
International Journal of Control, Automation, and Systems 2023; 21(11): 3746-3756