Back

Haptic Intelligence Members Publications

Learning Upper-Limb Exercises from Demonstrations

Final
(a) A user sitting in front of the PR2 robot. They each use one hand to hold an object that was specifically designed for upper-limb exercises for patients with stroke. Example trajectories from the (b) 1D, (c) 2D, and (d) pick-and-place exercises tested in the study.

Members

Thumb ticker sm siyaohu
Haptic Intelligence
Thumb ticker sm thumb ticker kjk 2024
Haptic Intelligence
Director

Publications

Haptic Intelligence Article Robotics for Occupational Therapy: Learning Upper-Limb Exercises From Demonstrations Hu, S., Mendonca, R., Johnson, M. J., Kuchenbecker, K. J. IEEE Robotics and Automation Letters, 6(4):7781-7788, October 2021 (Published)
We describe a learning-from-demonstration technique that enables a general-purpose humanoid robot to lead a user through object-mediated upper-limb exercises. It needs only tens of seconds of training data from a therapist teleoperating the robot to do the task with the user. We model the robot behavior as a regression problem, inferring the desired robot effort using the end-effector's state (position and velocity). Compared to the conventional approach of learning time-based trajectories, our strategy produces customized robot behavior and eliminates the need to tune gains to adapt to the user's motor ability. In our study, one occupational therapist and six people with stroke trained a Willow Garage PR2 on three example tasks (periodic 1D and 2D motions plus episodic pick-and-place). They then repeatedly did the tasks with the robot and blindly compared the state- and time-based controllers learned from the training data. Our results show that working models were reliably obtained to allow the robot to do the exercise with the user; that our state-based approach enabled users to be more actively involved, allowed larger excursion, and generated power outputs more similar to the therapist demonstrations; and that the therapist found our strategy more agreeable than the traditional time-based approach.
DOI BibTeX

Haptic Intelligence Article Hierarchical Task-Parameterized Learning from Demonstration for Collaborative Object Movement Hu, S., Kuchenbecker, K. J. Applied Bionics and Biomechanics, 2019(9765383):1-25, December 2019 (Published)
Learning from demonstration (LfD) enables a robot to emulate natural human movement instead of merely executing preprogrammed behaviors. This article presents a hierarchical LfD structure of task-parameterized models for object movement tasks, which are ubiquitous in everyday life and could benefit from robotic support. Our approach uses the task-parameterized Gaussian mixture model (TP-GMM) algorithm to encode sets of demonstrations in separate models that each correspond to a different task situation. The robot then maximizes its expected performance in a new situation by either selecting a good existing model or requesting new demonstrations. Compared to a standard implementation that encodes all demonstrations together for all test situations, the proposed approach offers four advantages. First, a simply defined distance function can be used to estimate test performance by calculating the similarity between a test situation and the existing models. Second, the proposed approach can improve generalization, e.g., better satisfying the demonstrated task constraints and speeding up task execution. Third, because the hierarchical structure encodes each demonstrated situation individually, a wider range of task situations can be modeled in the same framework without deteriorating performance. Last, adding or removing demonstrations incurs low computational load, and thus, the robot’s skill library can be built incrementally. We first instantiate the proposed approach in a simulated task to validate these advantages. We then show that the advantages transfer to real hardware for a task where naive participants collaborated with a Willow Garage PR2 robot to move a handheld object. For most tested scenarios, our hierarchical method achieved significantly better task performance and subjective ratings than both a passive model with only gravity compensation and a single TP-GMM encoding all demonstrations.
DOI BibTeX