-
Notifications
You must be signed in to change notification settings - Fork 9
Description
I followed the README instructions to set up the environment on a fresh machine (without uv cache), but encountered an error when running scripts/train/psi0/finetune-simple-psi0.sh:
Resolving data files: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 80/80 [00:00<00:00, 16376[90/1840]
[rank1]: Traceback (most recent call last):
[rank1]: File "/data3/zixi/Psi0/scripts/train.py", line 364, in
[rank1]: train(config) # type: ignore
[rank1]: File "/data3/zixi/Psi0/scripts/train.py", line 191, in train
[rank1]: train_dataset, val_dataset = trainer.create_datasets()
[rank1]: File "/data3/zixi/Psi0/src/psi/trainers/finetune.py", line 226, in create_datasets
[rank1]: self.train_dataset = self.data_cfg(split="train", transform_kwargs=transform_kwargs)
[rank1]: File "/data3/zixi/Psi0/src/psi/config/data_lerobot.py", line 45, in call
[rank1]: train_dataset = LeRobotDatasetWrapper(self, split=split)
[rank1]: File "/data3/zixi/Psi0/src/psi/data/lerobot/lerobot_ext.py", line 39, in init
[rank1]: self.base_dataset = lerobot_dataset_class(
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/lerobot/datasets/lerobot_dataset.py", line 480, in init
[rank1]: self.hf_dataset = self.load_hf_dataset()
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/lerobot/datasets/lerobot_dataset.py", line 601, in load_hf_dataset
[rank1]: hf_dataset = load_dataset("parquet", data_dir=path, split="train")
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/load.py", line 2062, in load_dataset
[rank1]: builder_instance = load_dataset_builder(
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/load.py", line 1819, in load_dataset_builder
[rank1]: builder_instance: DatasetBuilder = builder_cls(
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/builder.py", line 395, in init
[rank1]: self.info = DatasetInfo.from_directory(self._cache_dir)
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/info.py", line 279, in from_directory
[rank1]: return cls.from_dict(dataset_info_dict)
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/info.py", line 284, in from_dict
[rank1]: return cls(**{k: v for k, v in dataset_info_dict.items() if k in field_names})
[rank1]: File "", line 20, in init
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/info.py", line 170, in post_init
[rank1]: self.features = Features.from_dict(self.features)
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/features/features.py", line 1888, in from_dict
[rank1]: obj = generate_from_dict(dic)
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/features/features.py", line 1468, in generate_from_dict
[rank1]: return {key: generate_from_dict(value) for key, value in obj.items()}
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/features/features.py", line 1468, in
[rank1]: return {key: generate_from_dict(value) for key, value in obj.items()}
[rank1]: File "/data3/zixi/Psi0/.venv-psi/lib/python3.10/site-packages/datasets/features/features.py", line 1474, in generate_from_dict
[rank1]: raise ValueError(f"Feature type '{_type}' not found. Available feature types: {list(_FEATURE_TYPES.keys())}")
[rank1]: ValueError: Feature type 'List' not found. Available feature types: ['Value', 'ClassLabel', 'Translation', 'TranslationVariableLanguages', 'LargeList', 'Sequence', 'Array2D', 'Array3D', 'Array4D', 'Array5D', 'Audio', 'Image', 'Video', 'Pdf', 'VideoFrame']
This error seems to be related to the datasets version not being 4.0.0. However, when I upgrade to datasets==4.0.0, the lerobot library becomes incompatible. It appears that the lerobot version and data format used by Psi0 do not support directly upgrading datasets to 4.0.0.
Is there any way to avoid this error? Or has anyone encountered a similar issue during reproduction? Looking forward to your responses.