Affiliation:
1. Shizuoka Institute of Science and Technology, Shizuoka, Japan
Abstract
This research aims to enhance the collaborative work system between humans and robots by exploring “ensemble music.” In an ensemble, it is crucial to adhere to the score, synchronize it with the breathing of fellow musicians, and ensure harmonious performance. This represents one of the most intricate collaborative endeavors achievable by humans. In this study, by examining various image processing methods for detecting the movement of a performer, it was shown that skeleton detection using MediaPipe is appropriate in terms of a large amount of information and processing speed. Next, a deep neural network utilizing the history of MediaPipe’s 3D skeletal coordinates as the input was developed to detect the performance start and end points. A comprehensive examination of learning and estimation conditions via grid search revealed that the start and end points could be estimated with approximately 70% and 100% accuracy, respectively, when using a history of 10 points, the ReLU activation function, and the L-BFGS optimizer. Additionally, the estimation time was 10 ms or less when the hidden layer had 100 or fewer units. Future detection accuracy will be enhanced by incorporating additional learning data and assigning greater weights to skeleton points with significant changes.
Reference19 articles.
1. UNIVERSAL ROBOTS [Internet], [cited 2024 Aug 20]. Available from: https://www.universal-robots.com/applications/machine-tending/.
2. ISO 10218-1: 2011. Robots and robotic devices Safety requirements for industrial robots Part 1: Robots, 2011.
3. ISO 10218-2: 2011. Robots and robotic devices Safety requirements for industrial robots Part 2: Robot systems and integration, 2011.
4. ISO/TS 15066:2016. Robots and robotic devices Collaborative robots, 2016.
5. Ichinose S, Mizuno S, Shiramatsu S, Kitahara T. Two approaches to supporting improvisational ensemble for music beginners based on body motion tracking. Int J Smart Comput Artif Intel. 2019;3(1):55–70.