Regular Papers

International Journal of Control, Automation and Systems 2022; 20(10): 3410-3417

Published online August 27, 2022

https://doi.org/10.1007/s12555-021-0449-6

© The International Journal of Control, Automation, and Systems

Grasping Time and Pose Selection for Robotic Prosthetic Hand Control Using Deep Learning Based Object Detection

Hae-June Park, Bo-Hyeon An, Su-Bin Joo , Oh-Won Kwon, Min Young Kim*, and Joonho Seo*

Kyungpook National University and KIMM

Abstract

This paper presents an algorithm to control a robotic prosthetic hand by applying deep learning (DL) to select a grasping pose and a grasping time from 2D images and 3D point clouds. This algorithm consists of four steps: 1) Acquisition of 2D images and 3D point clouds of objects; 2) Object recognition in the 2D images; 3) Grasping pose selection; 4) Choice of a grasping time and control of the prosthetic hand. The grasping pose selection is necessary when the algorithm detects many objects in the same frame, and must decide which pose of the prosthetic hand should use. The pose was chosen considering the object that was to the prosthesis. The grasping time was determined by the operating point when approaching the selected target after selecting the grasping pose; this choice uses an empirically-determined distance threshold. The proposed method achieved 89% accuracy of the grasping the intended object. The failures occurred because of slight inaccuracy in object localization, occlusion of target objects, and the inability of DL object detection. Work to solve these shortcomings is ongoing. This algorithm will help to improve the convenience of the user of a prosthetic hand.

Keywords Computer vision, grasping pose selection, grasping time selection, point cloud, 3D distance.

Article

Regular Papers

International Journal of Control, Automation and Systems 2022; 20(10): 3410-3417

Published online October 1, 2022 https://doi.org/10.1007/s12555-021-0449-6

Copyright © The International Journal of Control, Automation, and Systems.

Grasping Time and Pose Selection for Robotic Prosthetic Hand Control Using Deep Learning Based Object Detection

Hae-June Park, Bo-Hyeon An, Su-Bin Joo , Oh-Won Kwon, Min Young Kim*, and Joonho Seo*

Kyungpook National University and KIMM

Abstract

This paper presents an algorithm to control a robotic prosthetic hand by applying deep learning (DL) to select a grasping pose and a grasping time from 2D images and 3D point clouds. This algorithm consists of four steps: 1) Acquisition of 2D images and 3D point clouds of objects; 2) Object recognition in the 2D images; 3) Grasping pose selection; 4) Choice of a grasping time and control of the prosthetic hand. The grasping pose selection is necessary when the algorithm detects many objects in the same frame, and must decide which pose of the prosthetic hand should use. The pose was chosen considering the object that was to the prosthesis. The grasping time was determined by the operating point when approaching the selected target after selecting the grasping pose; this choice uses an empirically-determined distance threshold. The proposed method achieved 89% accuracy of the grasping the intended object. The failures occurred because of slight inaccuracy in object localization, occlusion of target objects, and the inability of DL object detection. Work to solve these shortcomings is ongoing. This algorithm will help to improve the convenience of the user of a prosthetic hand.

Keywords: Computer vision, grasping pose selection, grasping time selection, point cloud, 3D distance.

IJCAS
January 2025

Vol. 23, No. 1, pp. 1~88

Stats or Metrics

Share this article on

  • line

Related articles in IJCAS

IJCAS

eISSN 2005-4092
pISSN 1598-6446