ccop's picture
Update README.md
f43bcbc verified
|
raw
history blame
5.83 kB
---
task_categories:
- robotics
tags:
- LeRobot
- RoboGrasp2024
- Virtual
pretty_name: RoboGrasp 2024 Hackathon Dataset
---
# 2024 RoboGrasp Dataset
This is dataset the dataset for the RoboGrasp Hackathon 2024.
It includes 108 simulated pick-and-place robot episodes collected on a [simulated mobile aloha environment](https://github.com/HumanoidTeam/gym-aloha-hackathon.git) through teleoperated demonstrations.
During the collected episodes the right arm of the robot is used to pick an item from the table and put it in a box on the top of the table.
There are three types of items:
- green cube (47% of episodes)
- red sphere (30% of episodes)
- blue cylinder (22% of episodes)
You can visualize the dataset episodes [here](https://huggingface.co/spaces/lerobot/visualize_dataset?dataset=HumanoidTeam%2Frobograsp_hackathon_2024&episode=0)
# Hackathon Task
## Clone Repository
As a first step for the hackathon: clone the [hackathon repository](https://github.com/HumanoidTeam/lerobot-hackathon.git) on the hackathon branch on your system.
It's a fork of the original LeRobot repository containing the assets necessary for running following tasks.
```bash
git clone -b hackathon https://github.com/HumanoidTeam/lerobot-hackathon.git
```
## Install Dependencies
After cloning the repository you can proceed in installing the dependencies using poetry. you can install poetry by running `pip install poetry`.
```bash
cd lerobot-hackathon
poetry lock
poetry build
poetry install --extras "gym-aloha"
```
This will create the virtual environment with all the dependencies required. You can source the environment by running `poetry shell` within the folder.
From this point, you can configure and run your policy training using all the models present in Lerobot (e.g. ACT, DiffusionPolicy, VQ-Bet, etc...).
## Policy Configuration
You can create a yaml file within the folder `lerobot/configs/policy/`. For example `robograsp2024_submission_model.yaml`.
Within the the yaml file you can configure the input/output data shapes, data normalization strategies, context length and policy parameters.
Some parts of the yaml file will be dependant on this dataset. Thus, we provide the parameters necessary to use this dataset.
Here are working configurations for the input and output structure:
These go at the beginning of the yaml:
```yaml
seed: 100000
dataset_repo_id: HumanoidTeam/robograsp_hackathon_2024
override_dataset_stats:
observation.images.left_wrist:
# stats from imagenet, since we use a pretrained vision model
mean: [[[0.485]], [[0.456]], [[0.406]]] # (c,1,1)
std: [[[0.229]], [[0.224]], [[0.225]]] # (c,1,1)
observation.images.right_wrist:
# stats from imagenet, since we use a pretrained vision model
mean: [ [ [ 0.485 ] ], [ [ 0.456 ] ], [ [ 0.406 ] ] ] # (c,1,1)
std: [ [ [ 0.229 ] ], [ [ 0.224 ] ], [ [ 0.225 ] ] ] # (c,1,1)
observation.images.top:
# stats from imagenet, since we use a pretrained vision model
mean: [[[0.485]], [[0.456]], [[0.406]]] # (c,1,1)
std: [[[0.229]], [[0.224]], [[0.225]]] # (c,1,1)
```
These go within the `policy:` scope and regards the input/output datashape
```yaml
input_shapes:
observation.images.left_wrist: [3, 480, 640]
observation.images.right_wrist: [3, 480, 640]
observation.images.top: [3, 480, 640]
observation.state: ["${env.state_dim}"]
output_shapes:
action: ["${env.action_dim}"]
# Normalization / Unnormalization
input_normalization_modes:
observation.images.left_wrist: mean_std
observation.images.right_wrist: mean_std
observation.images.top: mean_std
observation.state: min_max
output_normalization_modes:
action: min_max
```
The remaining configuration can be derived from the other examples provided by the lerobot original repo.
## Start Policy Training
you can start the policy training by running the following command, while having sourced the environment built in the previous section.
To source the environment run:
```bash
poetry shell
```
To start the training, you can use this command:
```bash
MUJOCO_GL="egl" python lerobot/scripts/train.py \
policy=robograsp2024_submission_model \
env=humanoid_hackathon_mobile_aloha \
env.task=AlohaHackathon-v0 \
dataset_repo_id=HumanoidTeam/robograsp_hackathon_2024
```
Where `robograsp2024_submission_model` is the name of the yaml file with the policy configuration, `humanoid_hackathon_mobile_aloha` is the provided yaml configuration for the mujoco environment to test the trained policies.
### Resume policy training from a checkpoint.
Terminated training too early? No worries! you can resume training from a previous checkpoint by running:
```bash
MUJOCO_GL="egl" python lerobot/scripts/train.py \
policy=robograsp2024_submission_model \
env=humanoid_hackathon_mobile_aloha \
env.task=AlohaHackathon-v0 \
dataset_repo_id=HumanoidTeam/robograsp_hackathon_2024 \
hydra.run.dir=OUTPUT_PATH \
resume=true
```
Where `OUTPUT_PATH` is the path to the checkpoint folder. It should look something like `outputs/train/2024-10-23/18-38-31_aloha_MODELTYPE_default`
## Upload trained policy checkpoint
After training the model you can upload it to Huggingface with:
```bash
huggingface-cli upload $hf_username/$repo_name PATH_TO_CHECKPOINT
```
where PATH_TO_CHECKPOINT is the folder containing the checkpoints of your training. it should look like `outputs/train/2024-10-23/23-02-55_aloha_diffusion_default/checkpoints/015000`.
# Policy Evaluation
```
python hackathon/evaluate_pretrained_policy_hackathon.py --device cuda --pretrained-policy-name-or-path HumanoidTeam/hackathon_sim_aloha --num-videos 5 --num-rollouts 10
```
This dataset was created using [🤗 LeRobot](https://github.com/huggingface/lerobot).