A person with paralysis cannot scratch their own itches. For most people this is trivial. For more than five million Americans living with paralysis, according to the RCHI Lab at Carnegie Mellon University, it is a daily indignity. Training a robot to do it requires data. Hiring five million people to demonstrate scratching is not an option. The fix, according to a paper released April 9, 2026 on arXiv, is generating the data synthetically.
The paper, from CMU's Robotic Caregiving and Human Interaction Lab under assistant professor Zackory Erickson, describes a pipeline that takes a text prompt like "scratch an itch on someone's left forearm" and produces a training dataset entirely in simulation. No human demonstrators required. The system uses Gemini 3 Pro to generate the scene: a soft-body human model with realistic anatomy, placed in a furnished room, with a robot motion plan to accomplish the task. It then generates four thousand variations of that scenario and trains a vision-based policy on the synthetic data. The result transfers to the real world without adjustment.
The robot used in the experiments is the Hello Robot Stretch 3, a mobile manipulator already deployed in research and home settings with disabled users, as documented by the Robot Report and Hello Robot's own community updates. Policies for scratching and bathing were trained entirely in simulation. When tested on real people making unscripted movements, the robot succeeded more than 80 percent of the time on both tasks.
What the pipeline actually generates, technically, is a chain of three systems working in sequence. Genesis, a physics simulator, handles the mechanics. ARCHITECT, a diffusion-based scene generator, produces the room layout and furniture. SMPL-X, a parameterized body model, creates an articulated soft human with twenty-seven degrees of freedom from ball joints at the shoulders, elbows, hips, knees, and neck. Gemini 3 Pro, the large language model from Google, serves as the controller: it takes the text prompt and coordinates all three systems, generating body pose parameters, selecting furniture placements, and producing robot motion as a sequence of end-effector waypoints. Full LLM and VLM prompts and example scenario specifications are on the RCHI Lab project page.
The critical constraint in physical human-robot interaction has always been data. Physical contact with a person is inherently safety-critical, which means any policy deployed in the real world must first be trained in simulation. But simulating a human being is not like simulating a box. Human tissue deforms. Bodies move unpredictably. A robot trained on one person's body shape may fail on another. The standard workaround is to collect physical demonstrations by teleoperating the robot, a research-assistant-intensive process that does not scale. This paper's claim is that the workaround is no longer necessary.
The paper reports an 80 percent success rate in a user study involving real humans. In simulation, the underlying trajectory generation succeeded 90 percent of the time for scratching and 70 percent of the time for bathing. The bathing number matters: the 30 percent of failed trajectories resulted from incorrect planner selection, insufficient contact time, or unintended contact with the wrong body part. When the robot is doing something delicate like washing someone, those failures are not acceptable. The authors note the gap as a known limitation.
The gap between two tasks and general caregiving is large. The pipeline can theoretically generate other activities of daily living, including wound dressing or feeding, by changing the text prompt. The paper does not test those tasks. Scaling from scratching and bathing to the full range of what a home caregiver does is an open problem.
Erickson's RCHI Lab has spent years studying how robots can physically assist people with disabilities. The Hello Robot Stretch 3 is an existing platform, not a research prototype, and Hello Robot has stated the long-term goal is the general-purpose home robot, not a dedicated care device. If the simulation data problem is solved, the path from lab to home becomes shorter. How much shorter depends on how far the text-prompt approach scales.
The paper is nine pages on arXiv and has not yet been peer reviewed. The 80 percent success figure comes from a user study, not a randomized controlled trial. The research was conducted with healthy participants, not people with disabilities, which is a meaningful distinction for the actual target population. These are honest caveats in the paper. The trajectory success rate in simulation for bathing being 30 percent lower than for scratching is also in the paper. What the paper does not answer is whether the synthetic training approach generalizes beyond the two tasks demonstrated.
The caregiving robot space has produced credible demos for years. What it has not produced is a scalable training method. This paper is an attempt at one.