logo DexTrack

Towards Generalizable Neural Tracking Control for Dexterous Manipulation
from Human References

1Tsinghua University, 2Shanghai Qi Zhi Institute, 3Shanghai AI Laboratory, 4UC San Diego
ICLR 2025

Our tracking controller can adeptly control a dexterous robot hand to accomplish various challenging manipulations involving new and challenging object geometries, novel and difficult manipulations with complex object movements as well as intricate and subtle in-hand re-orientations, and is robust to kinematics noise and out-of-domain (OOD) interactions involving objects from a brand new category.

Abstract

We address the challenge of developing a generalizable neural tracking controller for dexterous manipulation from human references. This controller aims to manage a dexterous robot hand to manipulate diverse objects for various purposes defined by kinematic human-object interactions. Developing such a controller is complicated by the intricate contact dynamics of dexterous manipulation and the need for adaptivity, generalizability, and robustness. Current reinforcement learning and trajectory optimization methods often fall short due to their dependence on task-specific rewards or precise system models. We introduce an approach that curates successful robot tracking demonstrations, comprising pairs of human references and robot actions, to train a neural controller. Utilizing a data flywheel, we iteratively enhance the controller's performance and the number of successful tracking demonstrations. We exploit available tracking demonstrations and carefully integrate reinforcement learning and imitation learning to boost the controller's performance in dynamic environments. At the same time, to obtain high-quality tracking demonstrations, we individually optimize per-trajectory tracking by leveraging the learned tracking controller in a homotopy optimization method. The homotopy optimization, mimicking chain-of-thought, aids in solving challenging trajectory tracking problems to increase demonstration diversity. We showcase our success by training a generalizable neural controller and evaluating it in both simulation and real world. Our method achieves over a 10% improvement in success rates compared to leading baselines.

Video

Tracking Control for Dexterous Manipulation

We can accurately track complex manipulations involving rich and changing contacts,
non-trivial object motions, and intricate tool-using.


Manipulating a thin shovel.
(Challenges: thin geoemtry, new object, novel interaction, complex object motions with intricate in-hand re-orientation)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Lifting and using a flashlight. (Challenges: thin object, novel interaction)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Interacting with a hand model. (Challenges: novel interaction, back-and-forth object movements)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Waving a extermely thin shovel. (Challenges: extermely thin geometry, novel interaction, complex object movememtns)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Using a short shovel. (Challenges: new object, novel interaction, complex object movememtns, intricate in-hand re-orientation)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Using a large water scoop. (Challenges: new object, complex object movememtns)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Using a short shovel. (Challenges: new object, novel interaction, complex object movememtns)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Lfiting a ball. (Challenges: novel interaction sequence, difficult object with round surface that is hard to grasp)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Waving a short shovel. (Challenges: new object, novel interaction, subtle in-hand re-orientation)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours


Lifting an extermely thin object. (Challenges: thin geoemtry, novel interaction)


Retargeted
Kinematic Reference

PPO
(tracking reward)

Ours

Real-World Evaluations

We transfer tracking results directly to the real world to assess tracking quality and also transfer the state-based controller to evaluate its robustness to noises in the state estimator.



Using a knife.

Using a hammer.

Using a soap.

Grasping and lifting a "hand".

Inspecting a small cube.

Grasping and passing a bottle.

Grasping and passing a mouse.

Grasping and passing a banana.

Grasping and using a watch.

Grasping and lifting an apple.

Using a phone.

Grasping and lifting a hammer.

Failure Cases

Effectiveness of the Homotopy Optimization Scheme

With the learned homotopy path generator, we can automatically plan valid optimization path for a dififuclt-to-track trajectory. By gradually solving each per-trajectory tracking problem in the path, we can finally get satisfactory results in originally unsolvable problems.



Retargeted
Kinematic Reference

w/o Homotopy Optimization

w/ Homotopy Optimization

Retargeted
Kinematic Reference

w/o Homotopy Optimization

w/ Homotopy Optimization

Retargeted
Kinematic Reference

w/o Homotopy Optimization

w/ Homotopy Optimization

Robustness towards Unreasoanble References and
Out-of-Distribution Interactions

Manage to track the full motion trajectory without being affected by large noise presented in the kinematics references containing unreasonable and unreachable motions


Kinematics References

Ours


Try to track the manipulation involving the object from a new objec category and unseen interaction triplets


Kinematics References

Ours

Failure Cases

Our method may fail to perform well in some cases where the object is from
a brand new category with challenging thin geometry.



Case 1

Case 2