i‐MYO: A multi‐grasp prosthetic hand control system based on gaze movements, augmented reality, and myoelectric signals
-
Published:2023-12-30
Issue:1
Volume:20
Page:
-
ISSN:1478-5951
-
Container-title:The International Journal of Medical Robotics and Computer Assisted Surgery
-
language:en
-
Short-container-title:Robotics Computer Surgery
Author:
Shi Chunyuan1ORCID,
Zhao Jingdong12,
Yang Dapeng12,
Jiang Li1
Affiliation:
1. State Key Laboratory of Robotics and System Harbin Institute of Technology (HIT) Harbin China
2. Artificial Intelligence Laboratory (HIT) Harbin China
Abstract
AbstractBackgroundControlling a multi‐grasp prosthetic hand still remains a challenge. This study explores the influence of merging gaze movements and augmented reality in bionics on improving prosthetic hand control.MethodsA control system based on gaze movements, augmented reality, and myoelectric signals (i‐MYO) was proposed. In the i‐MYO, the GazeButton was introduced into the controller to detect the grasp‐type intention from the eye‐tracking signals, and the proportional velocity scheme based on the i‐MYO was used to control hand movement.ResultsThe able‐bodied subjects with no prior training successfully transferred objects in 91.6% of the cases and switched the optimal grasp types in 97.5%. The patient could successfully trigger the EMG to control the hand holding the objects in 98.7% of trials in around 3.2 s and spend around 1.3 s switching the optimal grasp types in 99.2% of trials.ConclusionsMerging gaze movements and augmented reality in bionics can widen the control bandwidth of prosthetic hand. With the help of i‐MYO, the subjects can control a prosthetic hand using six grasp types if they can manipulate two muscle signals and gaze movement.
Funder
Postdoctoral Scientific Research Development Fund of Heilongjiang Province
Subject
Computer Science Applications,Biophysics,Surgery