Robot-Enhanced Therapies (RET) offer a promising alternative for Autism Spectrum Disorder (ASD) children. Within this framework, an imitation therapy is proposed where children replicate gestures demonstrated by a robot, including those derived from popular children songs. To enable effective feedback provision by the robot, gesture recognition becomes paramount. This paper introduces two approaches for achieving this goal: a multiclass classifier, intended to recognize which gesture is executed, and a set of binary classifiers discerning whether the expected gesture is performed or not. Both models rely on kinematic data acquired through the Azure Kinect camera and a Residual Network as classification model. Moreover, considering the challenges in children's data collection for model training, the work explores the impact that the collection of their data can bring to the outcomes of a gesture classification algorithm. Beside testing on adults and healthy children datasets, the study leverages a dataset comprising 69 gestures performed by ASD children during therapy sessions. Results indicate that binary models trained also on children data outperform the multiclass approach trained solely on adults data, achieving a median accuracy of 86%. This underscores the effectiveness of binary classifiers and suggests that integrating children's data can enhance algorithm performance. To strengthen findings, future research should expand dataset size, especially considering ASD children, and explore alternative action recognition algorithms like Long Short-Term Memories (LSTM).
Song Gesture Recognition for a Robot-Enhanced Imitation Therapy
Fassina Gabriele;Santos Laura.;Zorzella Elisa;Pedrocchi Alessandra;Ambrosini Emilia.
2024-01-01
Abstract
Robot-Enhanced Therapies (RET) offer a promising alternative for Autism Spectrum Disorder (ASD) children. Within this framework, an imitation therapy is proposed where children replicate gestures demonstrated by a robot, including those derived from popular children songs. To enable effective feedback provision by the robot, gesture recognition becomes paramount. This paper introduces two approaches for achieving this goal: a multiclass classifier, intended to recognize which gesture is executed, and a set of binary classifiers discerning whether the expected gesture is performed or not. Both models rely on kinematic data acquired through the Azure Kinect camera and a Residual Network as classification model. Moreover, considering the challenges in children's data collection for model training, the work explores the impact that the collection of their data can bring to the outcomes of a gesture classification algorithm. Beside testing on adults and healthy children datasets, the study leverages a dataset comprising 69 gestures performed by ASD children during therapy sessions. Results indicate that binary models trained also on children data outperform the multiclass approach trained solely on adults data, achieving a median accuracy of 86%. This underscores the effectiveness of binary classifiers and suggests that integrating children's data can enhance algorithm performance. To strengthen findings, future research should expand dataset size, especially considering ASD children, and explore alternative action recognition algorithms like Long Short-Term Memories (LSTM).File | Dimensione | Formato | |
---|---|---|---|
Fassina et al, 2024 - Song_Gesture_Recognition_for_a_Robot-Enhanced_Imitation_Therapy.pdf
Accesso riservato
Descrizione: Manuscript
:
Publisher’s version
Dimensione
3.98 MB
Formato
Adobe PDF
|
3.98 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.