The tests performed and the proposed methods work without changes on OAR and Grid5000 platforms, but with changes to the scripts, they can be applied on SLURM platforms.
There are different ways to run benchmarks in MLPerf, with Docker and Kadeploy being two of the most widely used tools due to their ease of integration and automation.
Docker allows you to package and distribute MLPerf test environments as portable containers (Rapid deployment), simplifying the installation of dependencies and ensuring consistent environments. This option is ideal for users seeking rapid configuration or working in infrastructures where container-based virtualization is standard. With Docker, you can obtain official MLCommons images or build custom containers tailored to specific hardware or software needs.
Kadeploy is a tool primarily used in research environments such as Grid'5000. It allows you to deploy custom operating systems on physical nodes in a cluster, granting full control over the test environment. This option is recommended when low-level access to the hardware is required or precise power consumption and performance measurements are desired without interference from the host system. Kadeploy allows you to run MLPerf benchmarks from scratch in a reproducible environment, without the overhead that virtualization can introduce.
Enter the Docker or Kadeploy directories to check out the different ways to run tests.