init commit
This commit is contained in:
176
README.md
176
README.md
@@ -1 +1,175 @@
|
||||
# unifolm-world-model-action
|
||||
# UnifoLM-WMA-0: A World-Model-Action (WMA) Framework under UnifoLM Family
|
||||
<p style="font-size: 1.2em;">
|
||||
<a href="https://unigen-x.github.io/unifolm-world-model-action.github.io"><strong>Project Page</strong></a> |
|
||||
<a href="https://https://huggingface.co/unitreerobotics/UnifoLM-WMA-0"><strong>Models</strong></a> |
|
||||
<a href="https://huggingface.co/unitreerobotics/datasets"><strong>Dataset</strong></a>
|
||||
</p>
|
||||
<div align="center">
|
||||
<p align="right">
|
||||
<span> 🌎English </span> | <a href="README_cn.md"> 🇨🇳中文 </a>
|
||||
</p>
|
||||
</div>
|
||||
<div align="justify">
|
||||
<b>UnifoLM-WMA-0</b> is Unitree‘s first open-source world-model–action architecture spanning multiple types of robotic embodiments, designed specifically for general-purpose robot learning. Its core component is a world-model capable of understanding the physical interactions between robots and the environments. This world-model provides two key functions: (a) <b>Simulation Engine</b> – operates as an interactive simulator to generate synthetic data for robot learning; (b) <b>Policy Enhancement</b> – connects with an action head and, by predicting future interaction processes with the world-model, further optimizes decision-making performance.
|
||||
</div>
|
||||
|
||||
## 🦾 Real Robot Deployment
|
||||
| <img src="assets/gifs/real_z1_stackbox.gif" style="border:none;box-shadow:none;margin:0;padding:0;" /> | <img src="assets/gifs/real_dual_stackbox.gif" style="border:none;box-shadow:none;margin:0;padding:0;" /> |
|
||||
|:---:|:---:|
|
||||
| <img src="assets/gifs/real_cleanup_pencils.gif" style="border:none;box-shadow:none;margin:0;padding:0;" /> | <img src="assets/gifs/real_g1_pack_camera.gif" style="border:none;box-shadow:none;margin:0;padding:0;" /> |
|
||||
|
||||
**Note: the top-right window shows the world model’s prediction of future environmental changes.**
|
||||
|
||||
## 📑 Opensource Plan
|
||||
- [x] Training
|
||||
- [x] Inference
|
||||
- [x] Checkpoints
|
||||
- [ ] Deployment
|
||||
|
||||
## ⚙️ Installation
|
||||
```
|
||||
conda create -n unifolm-wma python==3.10.14
|
||||
conda activate unifolma
|
||||
|
||||
conda install pinocchio=3.2.0 -c conda-forge -y
|
||||
conda install ffmpeg=7.1.1 -c conda-forge
|
||||
|
||||
git clone --recurse-submodules https://github.com/unitreerobotics/unifolm-world-model-action.git
|
||||
|
||||
# If you already downloaded the repo:
|
||||
git submodule update --init --recursive
|
||||
|
||||
cd unifolm-world-model-action
|
||||
pip install -e .
|
||||
|
||||
cd external/dlimp
|
||||
pip install -e .
|
||||
```
|
||||
## 🧰 Model Checkpoints
|
||||
| Model | Description | Link|
|
||||
|---------|-------|------|
|
||||
|$\text{UnifoLM-WMA-0}_{Base}$| Fintuned on [Open-X](https://robotics-transformer-x.github.io/) dataset. | [HuggingFace](https://huggingface.co/unitreerobotics/UnifoLM-WMA-0)|
|
||||
|$\text{UnifoLM-WMA-0}_{Dual}$| Fintuned on five [Unitree opensource dataset](https://huggingface.co/collections/unitreerobotics/g1-dex1-datasets-68bae98bf0a26d617f9983ab) in both decision-making and simulation modes. | [HuggingFace](https://huggingface.co/unitreerobotics/UnifoLM-WMA-0)|
|
||||
|
||||
## 🛢️ Dataset
|
||||
In our experiments, we consider the following three opensource dataset:
|
||||
| Dataset | Robot | Link |
|
||||
|---------|-------|------|
|
||||
|Z1_StackBox| [Unitree Z1](https://www.unitree.com/z1)|[Huggingface](https://huggingface.co/datasets/unitreerobotics/Z1_StackBox_Dataset)|
|
||||
|Z1_DualArm_StackBox|[Unitree Z1](https://www.unitree.com/z1)|[Huggingface](https://huggingface.co/datasets/unitreerobotics/Z1_DualArmStackBox_Dataset)|
|
||||
|Z1_DualArm_StackBox_V2|[Unitree Z1](https://www.unitree.com/z1)|[Huggingface](https://huggingface.co/datasets/unitreerobotics/Z1_DualArm_StackBox_Dataset_V2)|
|
||||
|Z1_DualArm_Cleanup_Pencils|[Unitree Z1](https://www.unitree.com/z1)|[Huggingface](https://huggingface.co/datasets/unitreerobotics/Z1_DualArm_CleanupPencils_Dataset)|
|
||||
|G1_Pack_Camera|[Unitree G1](https://www.unitree.com/g1)|[Huggingface](https://huggingface.co/datasets/unitreerobotics/G1_MountCameraRedGripper_Dataset)
|
||||
|
||||
To train on your own dataset, first to have the data following the [Huggingface LeRobot](https://github.com/huggingface/lerobot) dataset format. Assume the dataset’s source directory structure is as follows:
|
||||
```
|
||||
source_dir/
|
||||
├── dataset1_name
|
||||
├── dataset2_name
|
||||
├── dataset3_name
|
||||
└── ...
|
||||
```
|
||||
Then, convert a dataset to the required format using the command below:
|
||||
```python
|
||||
cd prepare_data
|
||||
python prepare_training_data.py \
|
||||
--source_dir /path/to/your/source_dir \
|
||||
--target_dir /path/to/save/the/converted/data \
|
||||
--dataset_name "dataset1_name" \
|
||||
--robot_name "a tag of the robot in the dataset" # e.g, Unitree Z1 Robot Arm or Unitree G1 Robot with Gripper.
|
||||
```
|
||||
The resulting data structure (Note: model training only supports input from the main-view camera. If the dataset includes multiple views, remove the corresponding values from the ```data_dir``` column in the CSV file.
|
||||
```
|
||||
target_dir/
|
||||
├── videos
|
||||
│ ├──dataset1_name
|
||||
│ │ ├──camera_view_dir
|
||||
│ │ ├── 0.mp4
|
||||
│ │ ├── 1.mp4
|
||||
│ │ └── ...
|
||||
│ └── ...
|
||||
├── transitions
|
||||
│ ├── dataset1_name
|
||||
│ ├── meta_data
|
||||
│ ├── 0.h5
|
||||
│ ├── 1.h5
|
||||
│ └── ...
|
||||
└── dataset1_name.csv
|
||||
```
|
||||
## 🚴♂️ Training
|
||||
To conduct training on a single or multiple datasets, please follow the steps below:
|
||||
- **Step 1**: The maximum DoF is assumed to be 16, if you have more than 16 DoF, update ```agent_state_dim``` and ```agent_action_dim``` in [configs/train/config.yaml](https://github.com/unitreerobotics/unifolm-wma/blob/working/configs/train/config.yaml) ;
|
||||
- **Step 2**: Set up the input shapes for each modality in [configs/train/meta.json](https://github.com/unitreerobotics/unitree-world-model/blob/main/configs/train/meta.json);
|
||||
- **Step 3**: Configure the training parameters in [configs/train/config.yaml](https://github.com/unitreerobotics/unitree-world-model/blob/main/configs/train/config.yaml). For the ```pretrained_checkpoint```, we recommend using the checkpoint " $\text{UnifoLM-WMA-0}_{Base}$ " fine-tuned on the [Open-X](https://robotics-transformer-x.github.io/) dataset;
|
||||
```yaml
|
||||
model:
|
||||
pretrained_checkpoint: /path/to/pretrained/checkpoint;
|
||||
...
|
||||
dicision_making_only: True # Train the world model only in decision-making mode. If False, jointly train it in both decision-making and simulation modes.
|
||||
...
|
||||
data:
|
||||
...
|
||||
train:
|
||||
...
|
||||
data_dir: /path/to/training/dataset/directory
|
||||
dataset_and_weights: # list the name of each dataset below and make sure the summation of weights is 1.0
|
||||
dataset1_name: 0.2
|
||||
dataset2_name: 0.2
|
||||
dataset3_name: 0.2
|
||||
dataset4_name: 0.2
|
||||
dataset5_name: 0.2
|
||||
```
|
||||
- **Step 4**: Setup ```experiment_name```, ```save_root``` variables in [scripts/train.sh](https://github.com/unitreerobotics/unitree-world-model/blob/main/scripts/train.sh);
|
||||
- **Step 5**: Lanuch the training with the command:
|
||||
```
|
||||
bash scripts/train.sh
|
||||
```
|
||||
## 🌏 Inference under the Interactive Simulation Mode
|
||||
To run the world model in an interactive simulation mode, follow these steps:
|
||||
- **Step 1**: (Skip this step if you just would like to test using the examples we provided) Prepare your own prompt following the format used in the [examples/world_model_interaction_prompts](https://github.com/unitreerobotics/unitree-world-model/tree/main/examples/world_model_interaction_prompts):
|
||||
```
|
||||
world_model_interaction_prompts/
|
||||
├── images
|
||||
│ ├── dataset1_name
|
||||
│ │ ├── 0.png # Image prompt
|
||||
│ │ └── ...
|
||||
│ └── ...
|
||||
├── transitions
|
||||
│ ├── dataset1_name
|
||||
│ │ ├── meta_data # Used for normalization
|
||||
│ │ ├── 0.h # Robot state and action data; in interaction mode,
|
||||
│ │ │ # only used to retrieve the robot state corresponding
|
||||
│ │ │ # to the image prompt
|
||||
│ │ └── ...
|
||||
│ └── ...
|
||||
├── dataset1_name.csv # File for loading image prompts, text instruction and corresponding robot states
|
||||
└── ...
|
||||
```
|
||||
- **Step 2**: Specify the correct paths for ```pretrained_checkpoint```(e.g, $\text{UnifoLM-WMA-0}_{Dual}$) and ```data_dir``` in [configs/inference/world_model_interaction.yaml](https://github.com/unitreerobotics/unitree-world-model/blob/main/configs/inference/world_model_interaction.yaml)
|
||||
- **Step 3**: Set the paths for ```checkpoint```, ```res_dir``` and ```prompt_dir``` in [scripts/run_world_model_interaction.sh](https://github.com/unitreerobotics/unitree-world-model/blob/main/scripts/run_world_model_interaction.sh), and specify all the dataset's name in ```datasets=(...)```. Then, lanuch the inference with the command:
|
||||
```
|
||||
bash scripts/run_world_model_interaction.sh
|
||||
```
|
||||
|
||||
## 📝 Codebase Architecture
|
||||
Here's a high-level overview of the project's code structure and core components:
|
||||
```
|
||||
unitree-world-model/
|
||||
├── assets # Media assets such as GIFs, images, and demo videos
|
||||
├── configs # Configuration files for training and inference
|
||||
│ ├── inference
|
||||
│ └── train
|
||||
├── examples # Example inputs and prompts for running inference
|
||||
├── external # External packages
|
||||
├── prepare_data # Scripts for dataset preprocessing and format conversion
|
||||
├── scripts # Main scripts for training, evaluation, and deployment
|
||||
├── src
|
||||
│ ├──unitree_worldmodel # Core Python package for the Unitree world model
|
||||
│ │ ├── data # Dataset loading, transformations, and dataloaders
|
||||
│ │ ├── models # Model architectures and backbone definitions
|
||||
│ │ ├── modules # Custom model modules and components
|
||||
│ │ └── utils # Utility functions and common helpers
|
||||
```
|
||||
|
||||
## 🙏 Acknowledgement
|
||||
Lots of code are inherieted from [DynamiCrafter](https://github.com/Doubiiu/DynamiCrafter), [Diffusion Policy](https://github.com/real-stanford/diffusion_policy) and [OpenVLA](https://github.com/openvla/openvla/tree/main).
|
||||
|
||||
Reference in New Issue
Block a user