66 pull_request :
77 branches : [ "main" ]
88
9- # env:
10- # CUDA_VISIBLE_DEVICES: 7
11-
129jobs :
1310 ci-temp-test-model :
1411 runs-on : ci-temp-test-model
1512 concurrency :
1613 group : ci-temp-test-model-${{ github.event.pull_request.number || github.ref }}
1714 cancel-in-progress : true
1815 steps :
16+ - name : Checkout accelerator-integration-wg
17+ uses : actions/checkout@v4
18+
1919 - name : Checkout flaggems
2020 uses : actions/checkout@v4
21+ with :
22+ repository : FlagOpen/FlagGems
23+ path : flag_gems
2124
2225 - name : Checkout benchmark
2326 uses : actions/checkout@v4
@@ -28,15 +31,16 @@ jobs:
2831 - name : Install and run benchmark
2932 shell : bash
3033 run : |
31- source tools/run_command.sh
32- run_command source "/root/anaconda3/etc/profile.d/conda.sh"
33- run_command conda init bash
34- run_command conda activate wz-torch-bench
35- run_command pip install -e .
36- run_command sed -i '/self\.worker\.run("import torch")/a\ self.worker.run(\
34+ # source tools/run_command.sh
35+ set -e
36+ source "/root/anaconda3/etc/profile.d/conda.sh"
37+ conda init bash
38+ conda activate new-torchbenchmark
39+ pip install -e flag_gems/
40+ sed -i '/self\.worker\.run("import torch")/a\ self.worker.run(\
3741 """\
3842 import flag_gems\
3943 flag_gems.enable(record=False, once=True, path='"'"'benchmark/oplist.log'"'"')""")' benchmark/torchbenchmark/__init__.py
40- run_command python benchmark/install.py models hf_GPT2
41- run_command python benchmark/run_benchmark.py test_bench --accuracy --device cuda --test eval --output output.json --models hf_GPT2
42- run_command sed -i '/self\.worker\.run($/,/^[[:space:]]*flag_gems\.enable.*oplist\.log.*""")/d' benchmark/torchbenchmark/__init__.py
44+ python benchmark/install.py models hf_GPT2
45+ python benchmark/run_benchmark.py test_bench --accuracy --device cuda --test eval --output output.json --models hf_GPT2
46+ sed -i '/self\.worker\.run($/,/^[[:space:]]*flag_gems\.enable.*oplist\.log.*""")/d' benchmark/torchbenchmark/__init__.py
0 commit comments