International Journal of Control, Automation and Systems 2022; 20(10): 3410-3417
Published online August 27, 2022
https://doi.org/10.1007/s12555-021-0449-6
© The International Journal of Control, Automation, and Systems
This paper presents an algorithm to control a robotic prosthetic hand by applying deep learning (DL) to select a grasping pose and a grasping time from 2D images and 3D point clouds. This algorithm consists of four steps: 1) Acquisition of 2D images and 3D point clouds of objects; 2) Object recognition in the 2D images; 3) Grasping pose selection; 4) Choice of a grasping time and control of the prosthetic hand. The grasping pose selection is necessary when the algorithm detects many objects in the same frame, and must decide which pose of the prosthetic hand should use. The pose was chosen considering the object that was to the prosthesis. The grasping time was determined by the operating point when approaching the selected target after selecting the grasping pose; this choice uses an empirically-determined distance threshold. The proposed method achieved 89% accuracy of the grasping the intended object. The failures occurred because of slight inaccuracy in object localization, occlusion of target objects, and the inability of DL object detection. Work to solve these shortcomings is ongoing. This algorithm will help to improve the convenience of the user of a prosthetic hand.
Keywords Computer vision, grasping pose selection, grasping time selection, point cloud, 3D distance.
International Journal of Control, Automation and Systems 2022; 20(10): 3410-3417
Published online October 1, 2022 https://doi.org/10.1007/s12555-021-0449-6
Copyright © The International Journal of Control, Automation, and Systems.
Hae-June Park, Bo-Hyeon An, Su-Bin Joo , Oh-Won Kwon, Min Young Kim*, and Joonho Seo*
Kyungpook National University and KIMM
This paper presents an algorithm to control a robotic prosthetic hand by applying deep learning (DL) to select a grasping pose and a grasping time from 2D images and 3D point clouds. This algorithm consists of four steps: 1) Acquisition of 2D images and 3D point clouds of objects; 2) Object recognition in the 2D images; 3) Grasping pose selection; 4) Choice of a grasping time and control of the prosthetic hand. The grasping pose selection is necessary when the algorithm detects many objects in the same frame, and must decide which pose of the prosthetic hand should use. The pose was chosen considering the object that was to the prosthesis. The grasping time was determined by the operating point when approaching the selected target after selecting the grasping pose; this choice uses an empirically-determined distance threshold. The proposed method achieved 89% accuracy of the grasping the intended object. The failures occurred because of slight inaccuracy in object localization, occlusion of target objects, and the inability of DL object detection. Work to solve these shortcomings is ongoing. This algorithm will help to improve the convenience of the user of a prosthetic hand.
Keywords: Computer vision, grasping pose selection, grasping time selection, point cloud, 3D distance.
Vol. 23, No. 1, pp. 1~88
Ru-Xiang Hua*, Hong-Xuan Ma, Wei Zou, Wei Zhang, and Zhuo Wang
International Journal of Control, Automation, and Systems 2024; 22(2): 615-630Jialong Xie, Botao Zhang*, Qiang Lu, and Oleg Borisov
International Journal of Control, Automation, and Systems 2024; 22(1): 252-264Hui Chen, Wanlou Chen, Yipeng Zuo, Peng Xu, and Zhonghua Hao*
International Journal of Control, Automation and Systems 2022; 20(10): 3321-3334