Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Dualsense Controller Teleoperation and ACT autonomy on ALOHA for Bigym benchmark tasks in mujoco

License

Notifications You must be signed in to change notification settings

rongweiji/Dualsense-ALOHA

 
 

Repository files navigation

Group 7

Dualsense-ALOHA

Dualsense™ Controller Teleoperation and ACT autonomy IL on ALOHA for one Bigym benchmark task

Outcomes:

  1. The ACT model successfully completed the task.
  2. Implemented a multi-buffer inference system for improved movement stability, and applied an inference result scaler to address action attenuation—ensuring that the movement amplitude was sufficient to avoid object penetration in this scenario 100% avoided.
  3. Integrated MuJoCo Warp in headless mode, achieving a 10x increase in step speed.

Intro:

  • Insipred by the Nitendo-Aloha
  • Reproduce the implementation and concept of Nintendo-Aloha and replace the controller to be compatible with Dualsense™ Controller
  • Explore ACT IL, Mujoco, Warp
  • Original Readme of bigym

Data collection Teleoprate based on Dualsense controller:

Teleoperation, refer to the ds_aloha.py script,based on pydualsense library.

Dishwasher Close Task GIF

Data.Collection.mp4

Inference :

Model training repository

Model inference entry point: controllers/demonstration/run_inference_temporalAgg.py

Policy.Infenrece.mp4

MuJoCo Warp Accelerator

Warp Compare

Reflections

1. Action Attenuation

After ACT model training, action outputs are often weaker than intended, leading to less accurate positioning (though movement direction is usually correct). This is likely due to most collected actions being near zero. To compensate, we apply a scaling factor to the inferred actions so the robot arm moves sufficiently. Comparing action distributions between the dataset and policy outputs is recommended.

Compare Data Collection and Inference
Policy.Execution.Scaler.compare.mp4

2. Discrete Action Handling

Binary controller actions (e.g., action2, action9) can become diluted during training. To address this, we use MixUp interpolation to convert discrete actions into smoother, continuous values, improving model learning and inference.

3. MuJoCo Warp Suitability

MuJoCo Warp is optimized for fast, headless multi-world simulations and is not ideal for single-world, visualized policy inference.

4. Others

Model training throug colab by A100 about 2s per it. vs my RTX3060 17s per it.

About

Dualsense Controller Teleoperation and ACT autonomy on ALOHA for Bigym benchmark tasks in mujoco

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 92.0%
  • Python 8.0%