Retargeted
Kinematic Reference
We address the challenge of developing a generalizable neural tracking controller for dexterous manipulation from human references. This controller aims to manage a dexterous robot hand to manipulate diverse objects for various purposes defined by kinematic human-object interactions. Developing such a controller is complicated by the intricate contact dynamics of dexterous manipulation and the need for adaptivity, generalizability, and robustness. Current reinforcement learning and trajectory optimization methods often fall short due to their dependence on task-specific rewards or precise system models. We introduce an approach that curates successful robot tracking demonstrations, comprising pairs of human references and robot actions, to train a neural controller. Utilizing a data flywheel, we iteratively enhance the controller's performance and the number of successful tracking demonstrations. We exploit available tracking demonstrations and carefully integrate reinforcement learning and imitation learning to boost the controller's performance in dynamic environments. At the same time, to obtain high-quality tracking demonstrations, we individually optimize per-trajectory tracking by leveraging the learned tracking controller in a homotopy optimization method. The homotopy optimization, mimicking chain-of-thought, aids in solving challenging trajectory tracking problems to increase demonstration diversity. We showcase our success by training a generalizable neural controller and evaluating it in both simulation and real world. Our method achieves over a 10% improvement in success rates compared to leading baselines.
Manipulating a thin shovel.
(Challenges: thin geoemtry, new object, novel interaction, complex object motions with intricate in-hand re-orientation)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Lifting and using a flashlight. (Challenges: thin object, novel interaction)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Interacting with a hand model. (Challenges: novel interaction, back-and-forth object movements)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Waving a extermely thin shovel. (Challenges: extermely thin geometry, novel interaction, complex object movememtns)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Using a short shovel. (Challenges: new object, novel interaction, complex object movememtns, intricate in-hand re-orientation)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Using a large water scoop. (Challenges: new object, complex object movememtns)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Using a short shovel. (Challenges: new object, novel interaction, complex object movememtns)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Lfiting a ball. (Challenges: novel interaction sequence, difficult object with round surface that is hard to grasp)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Waving a short shovel. (Challenges: new object, novel interaction, subtle in-hand re-orientation)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
Lifting an extermely thin object. (Challenges: thin geoemtry, novel interaction)
Retargeted
Kinematic Reference
PPO
(tracking reward)
Ours
We transfer tracking results directly to the real world to assess tracking quality and also transfer the state-based controller to evaluate its robustness to noises in the state estimator.
Using a knife.
Using a hammer.
Using a soap.
Grasping and lifting a "hand".
Inspecting a small cube.
Grasping and passing a bottle.
Grasping and passing a mouse.
Grasping and passing a banana.
Grasping and using a watch.
Grasping and lifting an apple.
Using a phone.
Grasping and lifting a hammer.
With the learned homotopy path generator, we can automatically plan valid optimization path for a dififuclt-to-track trajectory. By gradually solving each per-trajectory tracking problem in the path, we can finally get satisfactory results in originally unsolvable problems.
Retargeted
Kinematic Reference
w/o Homotopy Optimization
w/ Homotopy Optimization
Retargeted
Kinematic Reference
w/o Homotopy Optimization
w/ Homotopy Optimization
Retargeted
Kinematic Reference
w/o Homotopy Optimization
w/ Homotopy Optimization
Manage to track the full motion trajectory without being affected by large noise presented in the kinematics references containing unreasonable and unreachable motions
Kinematics References
Ours
Try to track the manipulation involving the object from a new objec category and unseen interaction triplets
Kinematics References
Ours
Our method may fail to perform well in some cases where the object is from
a brand new category with challenging thin geometry.
Case 1
Case 2