Cooperative Markov Decision Process model for human–machine co-adaptation in robot-assisted rehabilitation
- Publisher:
- ELSEVIER
- Publication Type:
- Journal Article
- Citation:
- Knowledge-Based Systems, 2024, 291
- Issue Date:
- 2024-05-12
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
1-s2.0-S0950705124002077-main.pdf | Published version | 741.24 kB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
Human–machine interaction is a critical component in robotic rehabilitation systems. A mutual learning strategy involving both machine- and human-oriented learning has shown improvements in learning efficiency and receptiveness. Despite these advancements, a theoretical framework that encompasses high-level human responses during robot-assisted rehabilitation is still needed. This paper introduces a novel human–machine interface that uses a Co-adaptive Markov Decision Process (CaMDP) model based on cooperative multi-agent reinforcement learning. The CaMDP model effectively measures user adaptation to machines, treating the entire rehabilitation process as a collaborative learning experience. It quantifies learning rates at a higher system abstraction level. Policy Iteration in Reinforcement Learning is employed for the cooperative adjustment of Policy Improvement between the human and machine. Simulation studies demonstrate that the proposed new Policy Improvement approach has great potential to address non-stationarity issues and significantly reduce the switching frequency of patients — from 16.6% to 2% in a sample of 120,000 cases. The CaMDP model provides valuable insights into rehabilitation effect prediction and risk avoidance through dual-agent simulation, thereby enhancing the overall performance of the rehabilitation process.
Please use this identifier to cite or link to this item: