OpenDriveLab

FreeTacMan

Robot-free Visuo-Tactile Data Collection System
for Contact-rich Manipulation

Longyan Wu1,4*  Checheng Yu2*  Jieji Ren3*  Li Chen2  
Ran Huang4  Guoying Gu3  Hongyang Li2,1
1SII
2HKU
3SJTU
4Fudan
arXiv 2025

Good things take time... and so does loading this page! Thanks for your patience! Best viewed in Chrome on a desktop.
FreeTacMan is a robot-free, human-centric visuo-tactile data collection system that enables the efficient transfer of human visual, tactile , and motor skills to robots. It facilitates the collection of large-scale, contact-rich manipulation datasets.

Enabling robots with contact-rich manipulation remains a pivotal challenge in robot learning, which is substantially hindered by the data collection gap, including its inefficiency and limited sensor setup.

Motivated by the dexterity and force feedback of human motion, we introduce FreeTacMan, a robot-free and human-centric visuo-tactile data collection system to acquire robot manipulation data accurately and efficiently. Our main contributions are:

1. An in-situ, robot-free, real-time tactile data-collection system that leverages a handheld gripper with modular visuo-tactile sensors to excel at diverse contact-rich tasks efficiently.

2. A large-scale, high-precision (sub-millimeter) visuo-tactile manipulation dataset with over 3000k visuo-tactile image pairs, more than 10k trajectories across 50 tasks.

3. Experimental validation shows that imitation policies trained with our visuo-tactile data achieve an average 50% higher success rate than vision-only approaches in a wide spectrum of contact-rich manipulation tasks.

Interactive Model Viewer

Dive into our 💡interactive 3D model viewer and explore the most popular native 3D formats with ease.
Try out the 🖱️move command to inspect internal structures.
It's more than just viewing — it's a hands-on exploration. Start 💫discovering now!

FreeTacMan on PIPERFreeTacMan on FRANKA

FreeTacMan features a universal gripper interface with quick-swap mounts compatible with various robots, such as Piper and Franka, with support for more platforms coming soon. It also includes a camera scaffold designed for precise alignment with the wrist-mounted camera, ensuring a consistent perspective. These components demonstrate the plug-and-play modularity of FreeTacMan, enabling seamless integration across diverse robotic platforms without requiring hardware-specific modifications.

Enabled by the efficient, precise, and fidelity tactile data collection system, we curate a diverse dataset of manipulation tasks spanning vision, touch, and proprioception modalities. The dataset spans 50 tasks, comprising more than 10k manipulation trajectories which contains over 3 million visuo-tactile image pairs.

Task dataset

We evaluate the effectiveness of FreeTacMan system and the quality of the dataset through a diverse set of contact-rich manipulation tasks. We integrate tactile feedback to assess its impact on policy performance, observing a substantial improvement that highlights its dynamic value in contact-rich tasks.Temporal-aware pretraining further enhances performance by aligning visual and tactile embeddings while capturing temporal dynamics. Across five evaluated tasks, imitation policies trained with our visuo-tactile data achieve an average success rate that is 50% higher than vision-only counterparts.

Policy Success Rate (%)
ACT (Vision-only)
Ours-α (Tactile)
Ours-β (Tactile Pretrained)

The robot grasps a plastic cup and places it stably on a tray without causing damage.

The videos are played at normal speed.

Policy Success Rate (%)
ACT: Vision-only
Ours-α: + Tactile
Ours-β: + Tactile Pretrained

We evaluate the usability of FreeTacMan through a user study involving 12 human participants with varying levels of experience, each collecting demonstrations across 5 tasks. Compared to previous setups, FreeTacMan consistently achieves the highest completion rates and efficiency, and is perceived as the most user-friendly and reliable data collection system.

CPUT Score
ALOHA
UMI
Ours

Fragile Cup

CPUT Score

USB Plug

CPUT Score

Texture Classification

CPUT Score

Stamp Press

CPUT Score

Calligraphy

CPUT Score

P.S.: Completion per Unit Time (CPUT), defined as completion_rate x efficiency

FreeTacMan © 2025