G1 (29DoF) + Dex3-1 |
H1_2 (Arm 7DoF) |
πRelease Note
-
Support IPC mode, defaulting to use SSHKeyboard for input control.
-
Merged motion mode support for H1_2 robot.
-
Merged motion mode support for the G1_23 robot arm.
-
Β·Β·Β·
This repository implements teleoperation control of a Unitree humanoid robot using XR (Extended Reality) devices (such as Apple Vision Pro, PICO 4 Ultra Enterprise, or Meta Quest 3).
If you have never worked with a Unitree robot before, please at least read up to the βApplication Developmentβ chapter in the official documentation first. Additionally, the Wiki of this repo contains a wealth of background knowledge that you can reference at any time.
Here are the required devices and wiring diagram,
The currently supported devices in this repository:
| π€ Robot | βͺ Status |
|---|---|
| G1 (29 DoF) | β Complete |
| G1 (23 DoF) | β Complete |
| H1 (4βDoF arm) | β Complete |
| H1_2 (7βDoF arm) | β Complete |
| Dex1β1 gripper | β Complete |
| Dex3β1 dexterous hand | β Complete |
| Inspire dexterous hand | β Complete |
| BrainCo dexterous hand | β Complete |
| Β·Β·Β· | Β·Β·Β· |
We tested our code on Ubuntu 20.04 and Ubuntu 22.04, other operating systems may be configured differently. This document primarily describes the default mode.
For more information, you can refer to Official Documentation and OpenTeleVision.
# Create a conda environment
(base) unitree@Host:~$ conda create -n tv python=3.10 pinocchio=3.1.0 numpy=1.26.4 -c conda-forge
(base) unitree@Host:~$ conda activate tv
# Clone this repo
(tv) unitree@Host:~$ git clone https://github.com/unitreerobotics/xr_teleoperate.git
(tv) unitree@Host:~$ cd xr_teleoperate
# Shallow clone submodule
(tv) unitree@Host:~/xr_teleoperate$ git submodule update --init --depth 1
# Install televuer submodule
(tv) unitree@Host:~/xr_teleoperate$ cd teleop/televuer
(tv) unitree@Host:~/xr_teleoperate/teleop/televuer$ pip install -e .
# Generate the certificate files required for televuer submodule
(tv) unitree@Host:~/xr_teleoperate/teleop/televuer$ openssl req -x509 -nodes -days 365 -newkey rsa:2048 -keyout key.pem -out cert.pem
# Install dex-retargeting submodule
(tv) unitree@Host:~/xr_teleoperate/teleop/televuer$ cd ../robot_control/dex-retargeting/
(tv) unitree@Host:~/xr_teleoperate/teleop/robot_control/dex-retargeting$ pip install -e .
# Install additional dependencies required by this repo
(tv) unitree@Host:~/xr_teleoperate/teleop/robot_control/dex-retargeting$ cd ../../../
(tv) unitree@Host:~/xr_teleoperate$ pip install -r requirements.txt# Install unitree_sdk2_python library which handles communication with the robot
(tv) unitree@Host:~$ git clone https://github.com/unitreerobotics/unitree_sdk2_python.git
(tv) unitree@Host:~$ cd unitree_sdk2_python
(tv) unitree@Host:~/unitree_sdk2_python$ pip install -e .Note 1: For
xr_teleoperateversions v1.1 and above, please ensure that theunitree_sdk2_pythonrepository is checked out to a commit equal to or newer than 404fe44d76f705c002c97e773276f2a8fefb57e4.Note 2: The unitree_dds_wrapper in the original h1_2 branch was a temporary version. It has now been fully migrated to the official Python-based control and communication library: unitree_sdk2_python.
Note 3: All identifiers in front of the command are meant for prompting: Which device and directory the command should be executed on.
In the Ubuntu system's
~/.bashrcfile, the default configuration is:PS1='${debian_chroot:+($debian_chroot)}\u@\h:\w\$ 'Taking the command
(tv) unitree@Host:~$ pip install meshcatas an example:
(tv)Indicates the shell is in the conda environment namedtv.unitree@Host:~Shows the user\uunitreeis logged into the device\hHost, with the current working directory\was$HOME.$shows the current shell is Bash (for non-root users).pip install meshcatis the commandunitreewants to execute onHost.You can refer to Harley Hahn's Guide to Unix and Linux and Conda User Guide to learn more.
First, install unitree_sim_isaaclab. Follow that repoβs README.
Then launch the simulation with a G1(29 DoF) and Dex3 hand configuration:
(base) unitree@Host:~$ conda activate unitree_sim_env
(unitree_sim_env) unitree@Host:~$ cd ~/unitree_sim_isaaclab
(unitree_sim_env) unitree@Host:~/unitree_sim_isaaclab$ python sim_main.py --device cpu --enable_cameras --task Isaac-PickPlace-Cylinder-G129-Dex3-Joint --enable_dex3_dds --robot_type g129π₯π₯π₯ NOTICEβ
After simulation starts, click once in the window to activate it.
The terminal will show:
controller started, start main loop...
Here is the simulation GUI:
This program supports XR control of a physical robot or in simulation. Choose modes with command-line arguments:
- Basic control parameters
| βοΈ Parameter | π Description | π Options | π Default |
|---|---|---|---|
--xr-mode |
Choose XR input mode | hand (hand tracking)controller (controller tracking) |
hand |
--arm |
Choose robot arm type (see 0. π Introduction) | G1_29G1_23H1_2H1 |
G1_29 |
--ee |
Choose end-effector (see 0. π Introduction) | dex1dex3inspire1brainco |
none |
- Mode flags
| βοΈ Flag | π Description |
|---|---|
--record |
Enable data recording After pressing r to start, press s to start/stop saving an episode. Can repeat. |
--motion |
Enable motion mode After enabling this mode, the teleoperation program can run alongside the robot's motion control. In hand tracking mode, you can use the R3 Controller to control the robot's walking behavior; in controller tracking mode, you can also use controllers to control the robotβs movement. |
--headless |
Run without GUI (for headless PC2 deployment) |
--sim |
Enable simulation mode |
Assuming hand tracking with G1(29 DoF) + Dex3 in simulation with recording:
(tv) unitree@Host:~$ cd ~/xr_teleoperate/teleop/
(tv) unitree@Host:~/xr_teleoperate/teleop/$ python teleop_hand_and_arm.py --xr-mode=hand --arm=G1_29 --ee=dex3 --sim --record
# Simplified (defaults apply):
(tv) unitree@Host:~/xr_teleoperate/teleop/$ python teleop_hand_and_arm.py --ee=dex3 --sim --recordAfter the program starts, the terminal shows:
Next steps:
-
Wear your XR headset (e.g. Apple Vision Pro, PICO4, etc.)
-
Connect to the corresponding WiβFi
-
Open a browser (e.g. Safari or PICO Browser) and go to:
https://192.168.123.2:8012?ws=wss://192.168.123.2:8012Note 1: This IP must match your Host IP (check with
ifconfig).Note 2: You may see a warning page. Click Advanced, then Proceed to IP (unsafe).
-
In the Vuer web, click Virtual Reality. Allow all prompts to start the VR session.
-
Youβll see the robotβs first-person view in the headset. The terminal prints connection info:
websocket is connected. id:dbb8537d-a58c-4c57-b49d-cbb91bd25b90 default socket worker is up, adding clientEvents Uplink task running. id:dbb8537d-a58c-4c57-b49d-cbb91bd25b90
-
Align your arm to the robotβs initial pose to avoid sudden movements at start:
-
Press r in the terminal to begin teleoperation. You can now control the robot arm and dexterous hand.
-
During teleoperation, press s to start recording; press s again to stop and save. Repeatable process.
Note 1: Recorded data is stored in
xr_teleoperate/teleop/utils/databy default, with usage instructions at this repo: unitree_IL_lerobot.Note 2: Please pay attention to your disk space size during data recording.
Press q in the terminal (or βrecord imageβ window) to quit.
Physical deployment steps are similar to simulation, with these key differences:
In the simulation environment, the image service is automatically enabled. For physical deployment, you need to manually start the image service based on your specific camera hardware. The steps are as follows:
Copy image_server.py in the xr_teleoperate/teleop/image_server directory to the Development Computing Unit PC2 of Unitree Robot (G1/H1/H1_2/etc.),
# p.s. You can transfer image_server.py to PC2 via the scp command and then use ssh to remotely login to PC2 to execute it.
# Assuming the IP address of the development computing unit PC2 is 192.168.123.164, the transmission process is as follows:
# log in to PC2 via SSH and create the folder for the image server
(tv) unitree@Host:~$ ssh unitree@192.168.123.164 "mkdir -p ~/image_server"
# Copy the local image_server.py to the ~/image_server directory on PC2
(tv) unitree@Host:~$ scp ~/xr_teleoperate/teleop/image_server/image_server.py unitree@192.168.123.164:~/image_server/and execute the following command in the PC2:
# p.s. Currently, this image transmission program supports two methods for reading images: OpenCV and Realsense SDK. Please refer to the comments in the `ImageServer` class within `image_server.py` to configure your image transmission service according to your camera hardware.
# Now located in Unitree Robot PC2 terminal
unitree@PC2:~/image_server$ python image_server.py
# You can see the terminal output as follows:
# {'fps': 30, 'head_camera_type': 'opencv', 'head_camera_image_shape': [480, 1280], 'head_camera_id_numbers': [0]}
# [Image Server] Head camera 0 resolution: 480.0 x 1280.0
# [Image Server] Image server has started, waiting for client connections...After image service is started, you can use image_client.py in the Host terminal to test whether the communication is successful:
(tv) unitree@Host:~/xr_teleoperate/teleop/image_server$ python image_client.pyNote 1: Skip this if your config does not use the Inspire hand.
Note 2: For G1 robot with Inspire DFX hand, related issue #46.
Note 3: For Inspire FTP hand, related issue #48.
First, use this URL: DFX_inspire_service to clone the dexterous hand control interface program. And Copy it to PC2 of Unitree robots.
On Unitree robot's PC2, execute command:
unitree@PC2:~$ sudo apt install libboost-all-dev libspdlog-dev
# Build project
unitree@PC2:~$ cd DFX_inspire_service && mkdir build && cd build
unitree@PC2:~/DFX_inspire_service/build$ cmake ..
unitree@PC2:~/DFX_inspire_service/build$ make -j6
# (For unitree g1) Terminal 1.
unitree@PC2:~/DFX_inspire_service/build$ sudo ./inspire_g1
# or (For unitree h1) Terminal 1.
unitree@PC2:~/DFX_inspire_service/build$ sudo ./inspire_h1 -s /dev/ttyUSB0
# Terminal 2. Run example
unitree@PC2:~/DFX_inspire_service/build$ ./hand_exampleIf two hands open and close continuously, it indicates success. Once successful, close the ./hand_example program in Terminal 2.
Please refer to the official documentation for setup instructions.
After installation, you need to manually start the services for both dexterous hands. Example commands are shown below (note: the serial port names may vary depending on your system):
# Terminal 1.
sudo ./brainco_hand --id 126 --serial /dev/ttyUSB1
# Terminal 2.
sudo ./brainco_hand --id 127 --serial /dev/ttyUSB2
- Everyone must keep a safe distance from the robot to prevent any potential danger!
- Please make sure to read the Official Documentation at least once before running this program.
- Without
--motion, always make sure that the robot has entered debug mode (L2+R2) to stop the motion control program, this will avoid potential command conflict problems.- To use motion mode (with
--motion), ensure the robot is in control mode (via R3 remote).- In motion mode:
- Right controller A = Exit teleop
- Both joysticks pressed = soft emergency stop (switch to damping mode)
- Left joystick = drive directions;
- right joystick = turning;
- max speed is limited in the code.
Same as simulation but follow the safety warnings above.
To avoid damaging the robot, it is recommended to position the robot's arms close to the initial pose before pressing q to exit.
In Debug Mode: After pressing the exit key, both arms will return to the robot's initial pose within 5 seconds, and then the control will end.
In Motion Mode: After pressing the exit key, both arms will return to the robot's motion control pose within 5 seconds, and then the control will end.
Same as simulation but follow the safety warnings above.
xr_teleoperate/
β
βββ assets [Storage of robot URDF-related files]
β
βββ hardware [3Dβprinted hardware modules]
β
βββ teleop
β βββ image_server
β β βββ image_client.py [Used to receive image data from the robot image server]
β β βββ image_server.py [Capture images from cameras and send via network (Running on robot's Development Computing Unit PC2)]
β β
β βββ televuer
β β βββ src/televuer
β β βββ television.py [captures XR devices's head, wrist, hand/controller data]
β β βββ tv_wrapper.py [Post-processing of captured data]
β β βββ test
β β βββ _test_television.py [test for television.py]
β β βββ _test_tv_wrapper.py [test for tv_wrapper.py]
β β
β βββ robot_control
β β βββ src/dex-retargeting [Dexterous hand retargeting algorithm library]
β β βββ robot_arm_ik.py [Inverse kinematics of the arm]
β β βββ robot_arm.py [Control dual arm joints and lock the others]
β β βββ hand_retargeting.py [Dexterous hand retargeting algorithm library Wrapper]
β β βββ robot_hand_inspire.py [Control inspire hand joints]
β β βββ robot_hand_unitree.py [Control unitree hand joints]
β β
β βββ utils
β β βββ episode_writer.py [Used to record data for imitation learning]
β β βββ weighted_moving_filter.py [For filtering joint data]
β β βββ rerun_visualizer.py [For visualizing data during recording]
β β
β βββ teleop_hand_and_arm.py [Startup execution code for teleoperation]
please see Device document.
This code builds upon following open-source code-bases. Please visit the URLs to see the respective LICENSES:
- https://github.com/OpenTeleVision/TeleVision
- https://github.com/dexsuite/dex-retargeting
- https://github.com/vuer-ai/vuer
- https://github.com/stack-of-tasks/pinocchio
- https://github.com/casadi/casadi
- https://github.com/meshcat-dev/meshcat-python
- https://github.com/zeromq/pyzmq
- https://github.com/Dingry/BunnyVisionPro
- https://github.com/unitreerobotics/unitree_sdk2_python