Skip to content

Commit e8a54a7

Browse files
ChonghaoSimaclaude
andcommitted
[UPDATE]: document advantage label release on HuggingFace & ModelScope
Update stage_advantage READMEs to note that Task_A/advantage/ is available on both HuggingFace and ModelScope dataset repos. Check off the HuggingFace & ModelScope to-do item and add update log entry. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
1 parent 9d350ca commit e8a54a7

File tree

3 files changed

+4
-3
lines changed

3 files changed

+4
-3
lines changed

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,7 @@ https://github.com/user-attachments/assets/3f5f0c48-ff3f-4b9b-985b-59ad0b2ea97c
5454

5555
## Update
5656

57+
- [Feb 15 2026] Stage Advantage **advantage labels** (`Task_A/advantage/`) released on [Hugging Face](https://huggingface.co/datasets/OpenDriveLab-org/Kai0) and [ModelScope](https://www.modelscope.cn/datasets/OpenDriveLab/Kai0).
5758
- [Feb 15 2026] Release of the **Train-Deploy Alignment** module: data augmentation (time scaling, space mirroring), DAgger data collection, inference with temporal smoothing/ensembling and RTC, and HDF5-to-LeRobot conversion.
5859
- [Feb 14 2026] Release of the **Stage Advantage** module: advantage estimator training, evaluation, GT labeling, and AWBC training pipeline.
5960
- [Feb 10 2026] Initial release of the **Model Arithmetic** module with support for both JAX and PyTorch checkpoints (not tested thoroughly).
@@ -212,7 +213,7 @@ Checkpoints are written to the config’s checkpoint directory. You can then use
212213
- [x] Model Arithmetic: code of different baselines for weight-space interpolation
213214
- [x] Stage Advantage: code, data (advantage labels), and checkpoints
214215
- [x] Train-Deploy Alignment: data augmentation, DAgger, inference (temporal smoothing, ensembling, RTC)
215-
- [ ] HuggingFace & ModelScope: upload Stage Advantage data and checkpoints
216+
- [x] HuggingFace & ModelScope: Stage Advantage data (`Task_A/advantage/`) and checkpoints uploaded
216217

217218
## Model Arithmetic
218219

stage_advantage/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ This module implements a pipeline for training an **Advantage Estimator** and us
2222

2323
**End-to-end order for AWBC:** (1) Stage 0 on data with `progress` → optional for Stage 1. (2) Stage 1 → train estimator. (3) Stage 2 → run eval on your dataset so it gets `data_PI06_100000/` or `data_KAI0_100000/` with advantage columns. (4) Run Stage 0 again with `--advantage-source absolute_advantage` on that dataset (e.g. via `gt_labeling.sh` with `DATA_PATH` = the repo you ran eval on, and source subdirs `data_PI06_100000` / `data_KAI0_100000`). (5) Point AWBC config `repo_id` at the resulting advantage-labeled directory and run Stage 3 training.
2424

25-
**Pre-annotated data:** The downloaded dataset includes **`data/Task_A/advantage`**, a fully annotated advantage dataset that can be used **directly for AWBC training** (Stage 3) without running Stage 0–2. Set the AWBC config `repo_id` to that path and run training.
25+
**Pre-annotated data:** The released dataset includes **`Task_A/advantage/`**, a fully annotated advantage dataset that can be used **directly for AWBC training** (Stage 3) without running Stage 0–2. It is available in both the [Hugging Face](https://huggingface.co/datasets/OpenDriveLab-org/Kai0) and [ModelScope](https://www.modelscope.cn/datasets/OpenDriveLab/Kai0) dataset repos. After downloading (e.g. via `scripts/download_dataset.py`), set the AWBC config `repo_id` to the local path (e.g. `<repo_root>/data/Task_A/advantage`) and run training.
2626

2727
---
2828

stage_advantage/awbc/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ Each uses `base_config=DataConfig(prompt_from_task=True)` so that the dataset’
1919
1. **Advantage dataset**
2020
The data must have `task_index` in each parquet and `meta/tasks.jsonl` (prompt strings per `task_index`).
2121

22-
**Pre-annotated data:** The downloaded dataset includes **`data/Task_A/advantage`**, a fully annotated advantage dataset that can be used **directly for AWBC training** (no need to run Stage 0–2 first). Set the AWBC config `repo_id` to that path and run the training commands below.
22+
**Pre-annotated data:** The released dataset includes **`Task_A/advantage/`**, a fully annotated advantage dataset that can be used **directly for AWBC training** (no need to run Stage 0–2 first). It is available in both the [Hugging Face](https://huggingface.co/datasets/OpenDriveLab-org/Kai0) and [ModelScope](https://www.modelscope.cn/datasets/OpenDriveLab/Kai0) dataset repos. After downloading, set the AWBC config `repo_id` to the local path (e.g. `<repo_root>/data/Task_A/advantage`) and run the training commands below.
2323

2424
To build your own advantage dataset instead:
2525
- Run **Stage 2** (eval) on your dataset → get `data_PI06_100000/` or `data_KAI0_100000/` with advantage columns.

0 commit comments

Comments
 (0)