Back to Search Start Over

Immersive virtual reality for learning exoskeleton-like virtual walking: a feasibility study.

Authors :
Rodríguez-Fernández A
den Berg AV
Cucinella SL
Lobo-Prat J
Font-Llagunes JM
Marchal-Crespo L
Source :
Journal of neuroengineering and rehabilitation [J Neuroeng Rehabil] 2024 Nov 01; Vol. 21 (1), pp. 195. Date of Electronic Publication: 2024 Nov 01.
Publication Year :
2024

Abstract

Purpose: Virtual Reality (VR) has proven to be an effective tool for motor (re)learning. Furthermore, with the current commercialization of low-cost head-mounted displays (HMDs), immersive virtual reality (IVR) has become a viable rehabilitation tool. Nonetheless, it is still an open question how immersive virtual environments should be designed to enhance motor learning, especially to support the learning of complex motor tasks. An example of such a complex task is triggering steps while wearing lower-limb exoskeletons as it requires the learning of several sub-tasks, e.g., shifting the weight from one leg to the other, keeping the trunk upright, and initiating steps. This study aims to find the necessary elements in VR to promote motor learning of complex virtual gait tasks.<br />Methods: In this study, we developed an HMD-IVR-based system for training to control wearable lower-limb exoskeletons for people with sensorimotor disorders. The system simulates a virtual walking task of an avatar resembling the sub-tasks needed to trigger steps with an exoskeleton. We ran an experiment with forty healthy participants to investigate the effects of first- (1PP) vs. third-person perspective (3PP) and the provision (or not) of concurrent visual feedback of participants' movements on the walking performance - namely number of steps, trunk inclination, and stride length -, as well as the effects on embodiment, usability, cybersickness, and perceived workload.<br />Results: We found that all participants learned to execute the virtual walking task. However, no clear interaction of perspective and visual feedback improved the learning of all sub-tasks concurrently. Instead, the key seems to lie in selecting the appropriate perspective and visual feedback for each sub-task. Notably, participants embodied the avatar across all training modalities with low cybersickness levels. Still, participants' cognitive load remained high, leading to marginally acceptable usability scores.<br />Conclusions: Our findings suggest that to maximize learning, users should train sub-tasks sequentially using the most suitable combination of person's perspective and visual feedback for each sub-task. This research offers valuable insights for future developments in IVR to support individuals with sensorimotor disorders in improving the learning of walking with wearable exoskeletons.<br /> (© 2024. The Author(s).)

Details

Language :
English
ISSN :
1743-0003
Volume :
21
Issue :
1
Database :
MEDLINE
Journal :
Journal of neuroengineering and rehabilitation
Publication Type :
Academic Journal
Accession number :
39487470
Full Text :
https://doi.org/10.1186/s12984-024-01482-y