Cooperative Markov Decision Process model for human–machine co-adaptation in robot-assisted rehabilitation

Publisher:
ELSEVIER
Publication Type:
Journal Article
Citation:
Knowledge-Based Systems, 2024, 291
Issue Date:
2024-05-12
Filename Description Size
1-s2.0-S0950705124002077-main.pdfPublished version741.24 kB
Adobe PDF
Full metadata record
Human–machine interaction is a critical component in robotic rehabilitation systems. A mutual learning strategy involving both machine- and human-oriented learning has shown improvements in learning efficiency and receptiveness. Despite these advancements, a theoretical framework that encompasses high-level human responses during robot-assisted rehabilitation is still needed. This paper introduces a novel human–machine interface that uses a Co-adaptive Markov Decision Process (CaMDP) model based on cooperative multi-agent reinforcement learning. The CaMDP model effectively measures user adaptation to machines, treating the entire rehabilitation process as a collaborative learning experience. It quantifies learning rates at a higher system abstraction level. Policy Iteration in Reinforcement Learning is employed for the cooperative adjustment of Policy Improvement between the human and machine. Simulation studies demonstrate that the proposed new Policy Improvement approach has great potential to address non-stationarity issues and significantly reduce the switching frequency of patients — from 16.6% to 2% in a sample of 120,000 cases. The CaMDP model provides valuable insights into rehabilitation effect prediction and risk avoidance through dual-agent simulation, thereby enhancing the overall performance of the rehabilitation process.
Please use this identifier to cite or link to this item: