Skip to content

Question about reproduction of Drivepi in open/closed-loop eval #17

@TabGuigui

Description

@TabGuigui

Thank you for sharing the excellent work. I have a few questions when reproducing the Drivepi baseline methods follow the instructions.

  1. Open-loop result.
    I reproduce the drivepi in this repo and evaluate the open-loop result as :
    [2025-12-31 02:50:16,310][src.agent.drivepi0.eval][INFO] - Loaded model from /data/DriveMoE/log/train/drive-pi0/2025-12-25_22-24_42/checkpoint/step2640.pt
    [2025-12-31 02:50:17,130][src.agent.drivepi0.eval][INFO] - load_checkpoint took 61.20 seconds to execute.
    [2025-12-31 02:50:24,090][src.agent.drivepi0.eval][INFO] - Using cuda device: cuda:0 dtype: torch.float32
    [2025-12-31 02:50:24,090][src.agent.drivepi0.eval][INFO] - Allocated GPU memory after loading model: 13.24 GB
    init took 0.98 seconds to execute.
    Agent done
    Running evaluation ...
    100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 385/385 [38:31<00:00, 6.00s/it
    ]
    [2025-12-31 03:28:58,468][src.agent.drivepi0.eval][INFO] - Number of episodes: 385
    [2025-12-31 03:28:58,468][src.agent.drivepi0.eval][INFO] - Eval | l1 Loss: 0.224 | l2 Loss: 0.187 | acc thres 0.05: 0.341 | acc thres 0.1: 0.630 | acc thres 0.2: 0.832 | acc thres 0.3: 0.906 | acc thres 0.5: 0.96
    2
    I use the checkpoint with step2640, and the average L2 loss is 0.187 on 385 episodes. However, this result is much lower than that reported in the paper(0.56), I don't know where I'm wrong.

  2. The sensor setting.
    I find in the training stage of Drivepi, only the front camera is used, while in the inference state of closed loop eval, I find in line 314 & 329 of drivepi0_carla_agent.py, it is default to use the additional back camera? I'm not sure why the setting is changed.

  3. Closed-loop time.
    I use the checkpoint I reproduce(since there is no provided checkpoint in this repo) in closed-loop eval, and the game time is much longer than that reported in the repo. Such as route RouteScenario_1825_rep, and the report game time is 139.4, while my checkpoint has to last over 190s. I wonder how long the evaluation will take for the whole 220 route for DrivePi

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions