Robot-Enhanced Therapies (RET) offer a promising alternative for Autism Spectrum Disorder (ASD) children. Within this framework, an imitation therapy is proposed where children replicate gestures demonstrated by a robot, including those derived from popular children songs. To enable effective feedback provision by the robot, gesture recognition becomes paramount. This paper introduces two approaches for achieving this goal: a multiclass classifier, intended to recognize which gesture is executed, and a set of binary classifiers discerning whether the expected gesture is performed or not. Both models rely on kinematic data acquired through the Azure Kinect camera and a Residual Network as classification model. Moreover, considering the challenges in children's data collection for model training, the work explores the impact that the collection of their data can bring to the outcomes of a gesture classification algorithm. Beside testing on adults and healthy children datasets, the study leverages a dataset comprising 69 gestures performed by ASD children during therapy sessions. Results indicate that binary models trained also on children data outperform the multiclass approach trained solely on adults data, achieving a median accuracy of 86%. This underscores the effectiveness of binary classifiers and suggests that integrating children's data can enhance algorithm performance. To strengthen findings, future research should expand dataset size, especially considering ASD children, and explore alternative action recognition algorithms like Long Short-Term Memories (LSTM).

Song Gesture Recognition for a Robot-Enhanced Imitation Therapy

Fassina Gabriele;Santos Laura.;Zorzella Elisa;Pedrocchi Alessandra;Ambrosini Emilia.
2024-01-01

Abstract

Robot-Enhanced Therapies (RET) offer a promising alternative for Autism Spectrum Disorder (ASD) children. Within this framework, an imitation therapy is proposed where children replicate gestures demonstrated by a robot, including those derived from popular children songs. To enable effective feedback provision by the robot, gesture recognition becomes paramount. This paper introduces two approaches for achieving this goal: a multiclass classifier, intended to recognize which gesture is executed, and a set of binary classifiers discerning whether the expected gesture is performed or not. Both models rely on kinematic data acquired through the Azure Kinect camera and a Residual Network as classification model. Moreover, considering the challenges in children's data collection for model training, the work explores the impact that the collection of their data can bring to the outcomes of a gesture classification algorithm. Beside testing on adults and healthy children datasets, the study leverages a dataset comprising 69 gestures performed by ASD children during therapy sessions. Results indicate that binary models trained also on children data outperform the multiclass approach trained solely on adults data, achieving a median accuracy of 86%. This underscores the effectiveness of binary classifiers and suggests that integrating children's data can enhance algorithm performance. To strengthen findings, future research should expand dataset size, especially considering ASD children, and explore alternative action recognition algorithms like Long Short-Term Memories (LSTM).
2024
IEEE International Workshop on Robot and Human Communication, RO-MAN
File in questo prodotto:
File Dimensione Formato  
Fassina et al, 2024 - Song_Gesture_Recognition_for_a_Robot-Enhanced_Imitation_Therapy.pdf

Accesso riservato

Descrizione: Manuscript
: Publisher’s version
Dimensione 3.98 MB
Formato Adobe PDF
3.98 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1287639
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact