Regular Papers

International Journal of Control, Automation and Systems 2022; 20(12): 3996-4004

Published online November 3, 2022

https://doi.org/10.1007/s12555-021-0929-8

© The International Journal of Control, Automation, and Systems

RANET: A Grasp Generative Residual Attention Network for Robotic Grasping Detection

Qian-Qian Hong, Liang Yang*, and Bi Zeng

University of Electronic Science and Technology of China Zhongshan Institute

Abstract

This paper presents a novel grasp generative residual attention network (RANET) for generating antipodal robotic grasp from multi-modal images with the pixel-wise method. To strengthen the generalization ability of unknown objects, this paper proposed a new structure that differs from the previous grasp generative network in that it additionally integrates a coordinate attention mechanism and a symmetrical skip connection, respectively. Using the coordinate attention module to emphasize meaningful information of the feature map and the symmetrical skip connection to remain more fine-grained details of feature. Moreover, a multi atrous convolution module is included in the structure to capture more high-level information, while a hypercolumn feature fusion method is incorporated for getting the best from the complementation of different layers’ features. Through evaluation on public datasets, the result demonstrates that we achieve 98.9% accuracy on the Cornell dataset which is the state-of-the-art performance with real-time speed(∼ 17 ms), meanwhile, we represent a 93.9% accuracy performance on the Jacquard dataset.

Keywords Convolutional neural networks, deep learning, grasping detection, vision.

Article

Regular Papers

International Journal of Control, Automation and Systems 2022; 20(12): 3996-4004

Published online December 1, 2022 https://doi.org/10.1007/s12555-021-0929-8

Copyright © The International Journal of Control, Automation, and Systems.

RANET: A Grasp Generative Residual Attention Network for Robotic Grasping Detection

Qian-Qian Hong, Liang Yang*, and Bi Zeng

University of Electronic Science and Technology of China Zhongshan Institute

Abstract

This paper presents a novel grasp generative residual attention network (RANET) for generating antipodal robotic grasp from multi-modal images with the pixel-wise method. To strengthen the generalization ability of unknown objects, this paper proposed a new structure that differs from the previous grasp generative network in that it additionally integrates a coordinate attention mechanism and a symmetrical skip connection, respectively. Using the coordinate attention module to emphasize meaningful information of the feature map and the symmetrical skip connection to remain more fine-grained details of feature. Moreover, a multi atrous convolution module is included in the structure to capture more high-level information, while a hypercolumn feature fusion method is incorporated for getting the best from the complementation of different layers’ features. Through evaluation on public datasets, the result demonstrates that we achieve 98.9% accuracy on the Cornell dataset which is the state-of-the-art performance with real-time speed(∼ 17 ms), meanwhile, we represent a 93.9% accuracy performance on the Jacquard dataset.

Keywords: Convolutional neural networks, deep learning, grasping detection, vision.

IJCAS
January 2025

Vol. 23, No. 1, pp. 1~88

Stats or Metrics

Share this article on

  • line

Related articles in IJCAS

IJCAS

eISSN 2005-4092
pISSN 1598-6446