Skip to content

OpenDriveLab/TAMEn

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 

Repository files navigation

TAMEn: Tactile-Aware Manipulation Engine for Closed-Loop Data Collection in Contact-Rich Tasks

🚀 Website | 📄 Paper | 🛠️ Hardware (Coming Soon)

✒️ Longyan Wu1,2,3, Jieji Ren4, Chenghang Jiang5, Junxi Zhou5, Shijia Peng3, Ran Huang1, Guoying Gu4, Li Chen3, Hongyang Li2,3

💼 1 Fudan University; 2 Shanghai Innovation Institute; 3 OpenDriveLab at The University of Hong Kong; 4 Shanghai Jiao Tong University; 5 East China University of Science and Technology

🦾 Highlight

  • (i) A visuo-tactile data engine for bimanual contact-rich manipulation, which integrates hardware, acquisition strategy, and policy learning into a closed-loop framework.
  • (ii) A human-machine interface that supports a dual-mode pipeline with sub-millimeter MoCap and VR-based in-the-wild acquisition, and can rapidly adapt to heterogeneous grippers.
  • (iii) A data collection recipe that incorporates real-time validation during collection and organizes heterogeneous multi-modal data into a pyramid-structured regime for staged learning.
  • (iv) A closed-loop data flywheel that leverages AR-based teleoperation with tactile feedback (tAmeR) to refine policies using corrective data from realistic failures.

🎯 Table of Contents

✅ TODO List

Note: TAMEn follows a staged release plan for hardware, data collection, and policy learning.

  • Release tAmeR app for Pico 4 Ultra / Pico 4
  • Release CAD models for multimodal data collection devices (compatible with GelSight, Xense, DW-Tac, PaXini, and our own sensor)
  • Release data collection workflow and dataset
  • Release training scripts, model configs, and inference pipeline

🎮 Getting Started

tAmeR Quick Start

tAmeR is an AR app for robot teleoperation, providing operators with real-time visual and tactile feedback.

  • tAmeR runs on PICO 4 / 4 Ultra and sends controller poses + button states to a PC via TCP.
  • It can be directly applied to teleoperation for arbitrary robot arms, and we will soon open-source our teleoperation programs for JAKA K1 and AgileX Piper.
  • A ROS2 WebSocket-JPEG backend can be used to stream multi-camera visual/tactile images.

1) Prerequisites

  • Headset side:
    • tAmeR.apk installed on PICO
    • PICO and PC are in the same LAN
    • Controllers are paired and tracked
  • PC side:
    • A TCP receiver is running on your PC
    • TCP port (default 8018) is open and matches the app config
    • Wrist cameras and visuo-tactile cameras are connected to the PC and publishing ROS2 image topics

2) Video Backend (ROS2 -> WebSocket JPEG)

Dependencies

sudo apt update
sudo apt install -y python3-opencv python3-aiohttp python3-numpy
source /opt/ros/<your_ros_distro>/setup.bash
python3 -c "import rclpy; from sensor_msgs.msg import Image; print('ros2 ok')"

Start backend service

# optional venv
# python3 -m venv tamen && source tamen/bin/activate && pip install --upgrade pip && pip install numpy opencv-python aiohttp

source /opt/ros/<your_ros_distro>/setup.bash
python3 tAmeR/tAmeR_ws.py \
  --host 0.0.0.0 \
  --port 8765 \
  --left-topic /left_camera/color/image_raw \
  --right-topic /right_camera/color/image_raw \
  --right-tactile-topic /right_tactile_camera/color/image_raw \
  --left-tactile-topic /left_tactile_camera/color/image_raw \
  --tile-width 320 \
  --tile-height 240 \
  --fps 10 \
  --jpeg-quality 70

3) Run VR Collection on Headset

  1. Launch tAmeR.apk on PICO.
  2. Input:
    • PC IP (LAN IP of your receiver machine)
    • Port (8018 by default, or your custom TCP port)
    • Period (0.01 recommended)
  3. Click Connect.
  4. After successful connection:
    • App starts sending data automatically
    • Input panel is hidden
    • Disconnect button is shown

4) Data Format (TCP)

Each line is semicolon-separated: timestamp;LG=T/F;RG=T/F;LT=T/F;RT=T/F;left_pose;right_pose;X=T/F;A=T/F;Y=T/F;B=T/F

  • LG/RG: grip buttons
  • LT/RT: triggers
  • X/Y/A/B: face buttons
  • left_pose/right_pose: x y z rx ry rz

📝 Citation

If you find this project useful in your research, please consider citing:

@misc{wu2026tamentactileawaremanipulationengine,
      title={TAMEn: Tactile-Aware Manipulation Engine for Closed-Loop Data Collection in Contact-Rich Tasks}, 
      author={Longyan Wu and Jieji Ren and Chenghang Jiang and Junxi Zhou and Shijia Peng and Ran Huang and Guoying Gu and Li Chen and Hongyang Li},
      year={2026},
      eprint={2604.07335},
      archivePrefix={arXiv},
      primaryClass={cs.RO},
      url={https://arxiv.org/abs/2604.07335}, 
}

🙏 Acknowledgements

We gratefully acknowledge Qianyu Guo, Checheng Yu, Chonghao Sima, Jingmin Zhang, and Chenyu Lin for their valuable insights and constructive discussions. We also extend our sincere gratitude to JAKA for their generous hardware and technical support.

📄 License

This project is licensed under the Apache License 2.0.

About

Tactile-Aware Manipulation Engine for Closed-Loop Data Collection in Contact-Rich Tasks

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages