|
1 | 1 | name: SYCL E2E |
2 | 2 |
|
3 | 3 | on: |
4 | | - workflow_call: |
5 | | - inputs: |
6 | | - name: |
7 | | - type: string |
8 | | - required: True |
| 4 | + # workflow_call: |
| 5 | + # inputs: |
| 6 | + # name: |
| 7 | + # type: string |
| 8 | + # required: True |
9 | 9 |
|
10 | | - runner: |
11 | | - type: string |
12 | | - required: True |
13 | | - image: |
14 | | - type: string |
15 | | - required: False |
16 | | - image_options: |
17 | | - type: string |
18 | | - required: True |
| 10 | + # runner: |
| 11 | + # type: string |
| 12 | + # required: True |
| 13 | + # image: |
| 14 | + # type: string |
| 15 | + # required: False |
| 16 | + # image_options: |
| 17 | + # type: string |
| 18 | + # required: True |
19 | 19 |
|
20 | | - target_devices: |
21 | | - type: string |
22 | | - required: False |
23 | | - extra_cmake_args: |
24 | | - type: string |
25 | | - required: False |
26 | | - tests_selector: |
27 | | - description: | |
28 | | - Three possible options: "e2e", "cts", and "benchmarks". |
29 | | - type: string |
30 | | - default: "e2e" |
| 20 | + # target_devices: |
| 21 | + # type: string |
| 22 | + # required: False |
| 23 | + # extra_cmake_args: |
| 24 | + # type: string |
| 25 | + # required: False |
| 26 | + # tests_selector: |
| 27 | + # description: | |
| 28 | + # Three possible options: "e2e", "cts", and "benchmarks". |
| 29 | + # type: string |
| 30 | + # default: "e2e" |
31 | 31 |
|
32 | | - extra_lit_opts: |
33 | | - description: | |
34 | | - Extra options to be added to LIT_OPTS. |
35 | | - type: string |
36 | | - default: '' |
| 32 | + # extra_lit_opts: |
| 33 | + # description: | |
| 34 | + # Extra options to be added to LIT_OPTS. |
| 35 | + # type: string |
| 36 | + # default: '' |
37 | 37 |
|
38 | | - repo_ref: |
39 | | - type: string |
40 | | - required: False |
41 | | - description: | |
42 | | - Commit SHA or branch to checkout the intel/llvm repo. |
43 | | - tests_ref: |
44 | | - type: string |
45 | | - required: False |
46 | | - description: Commit SHA or branch to checkout e2e/cts tests. |
| 38 | + # repo_ref: |
| 39 | + # type: string |
| 40 | + # required: False |
| 41 | + # description: | |
| 42 | + # Commit SHA or branch to checkout the intel/llvm repo. |
| 43 | + # tests_ref: |
| 44 | + # type: string |
| 45 | + # required: False |
| 46 | + # description: Commit SHA or branch to checkout e2e/cts tests. |
47 | 47 |
|
48 | | - toolchain_artifact: |
49 | | - type: string |
50 | | - default: '' |
51 | | - required: False |
52 | | - toolchain_artifact_filename: |
53 | | - type: string |
54 | | - default: '' |
55 | | - required: False |
56 | | - toolchain_decompress_command: |
57 | | - type: string |
58 | | - default: '' |
59 | | - required: False |
| 48 | + # toolchain_artifact: |
| 49 | + # type: string |
| 50 | + # default: '' |
| 51 | + # required: False |
| 52 | + # toolchain_artifact_filename: |
| 53 | + # type: string |
| 54 | + # default: '' |
| 55 | + # required: False |
| 56 | + # toolchain_decompress_command: |
| 57 | + # type: string |
| 58 | + # default: '' |
| 59 | + # required: False |
60 | 60 |
|
61 | | - binaries_artifact: |
62 | | - description: | |
63 | | - When set in modes other than `run-only` results in artifact upload. |
64 | | - For `run-only` mode, if specified, means downloading pre-built |
65 | | - binaries from the artifact. If set to special value `in-container` |
66 | | - then the binaries are expected to be stored in the container image |
67 | | - instead of coming from an artifact. |
68 | | - type: string |
69 | | - default: '' |
70 | | - required: False |
71 | | - testing_mode: |
72 | | - description: | |
73 | | - Testing mode to run E2E/CTS tests in, can be either `full`, `build-only` |
74 | | - or `run-only`. |
75 | | - type: string |
76 | | - default: 'full' |
77 | | - retention-days: |
78 | | - description: 'E2E/SYCL-CTS binaries artifact retention period.' |
79 | | - type: string |
80 | | - default: 1 |
| 61 | + # binaries_artifact: |
| 62 | + # description: | |
| 63 | + # When set in modes other than `run-only` results in artifact upload. |
| 64 | + # For `run-only` mode, if specified, means downloading pre-built |
| 65 | + # binaries from the artifact. If set to special value `in-container` |
| 66 | + # then the binaries are expected to be stored in the container image |
| 67 | + # instead of coming from an artifact. |
| 68 | + # type: string |
| 69 | + # default: '' |
| 70 | + # required: False |
| 71 | + # testing_mode: |
| 72 | + # description: | |
| 73 | + # Testing mode to run E2E/CTS tests in, can be either `full`, `build-only` |
| 74 | + # or `run-only`. |
| 75 | + # type: string |
| 76 | + # default: 'full' |
| 77 | + # retention-days: |
| 78 | + # description: 'E2E/SYCL-CTS binaries artifact retention period.' |
| 79 | + # type: string |
| 80 | + # default: 1 |
81 | 81 |
|
82 | | - install_igc_driver: |
83 | | - type: string |
84 | | - required: False |
85 | | - install_dev_igc_driver: |
86 | | - type: string |
87 | | - required: False |
88 | | - env: |
89 | | - type: string |
90 | | - default: '{}' |
91 | | - required: False |
| 82 | + # install_igc_driver: |
| 83 | + # type: string |
| 84 | + # required: False |
| 85 | + # install_dev_igc_driver: |
| 86 | + # type: string |
| 87 | + # required: False |
| 88 | + # env: |
| 89 | + # type: string |
| 90 | + # default: '{}' |
| 91 | + # required: False |
92 | 92 |
|
93 | | - skip_run: |
94 | | - type: string |
95 | | - default: 'false' |
96 | | - required: False |
| 93 | + # skip_run: |
| 94 | + # type: string |
| 95 | + # default: 'false' |
| 96 | + # required: False |
97 | 97 |
|
98 | | - benchmark_upload_results: |
99 | | - description: | |
100 | | - Set to true to upload results to git repository storing benchmarking |
101 | | - results. |
102 | | - type: string |
103 | | - default: 'false' |
104 | | - required: False |
105 | | - benchmark_save_name: |
106 | | - description: | |
107 | | - Save name to use for benchmark results: Save names are stored in |
108 | | - metadata of result file, and are used to identify benchmark results in |
109 | | - the same series (e.g. same configuration, same device, etc.). |
| 98 | + # benchmark_upload_results: |
| 99 | + # description: | |
| 100 | + # Set to true to upload results to git repository storing benchmarking |
| 101 | + # results. |
| 102 | + # type: string |
| 103 | + # default: 'false' |
| 104 | + # required: False |
| 105 | + # benchmark_save_name: |
| 106 | + # description: | |
| 107 | + # Save name to use for benchmark results: Save names are stored in |
| 108 | + # metadata of result file, and are used to identify benchmark results in |
| 109 | + # the same series (e.g. same configuration, same device, etc.). |
110 | 110 |
|
111 | | - Note: Currently, benchmark result filenames are in the format of |
112 | | - <benchmark_save_name>_<Device>_<Backend>_YYYYMMDD_HHMMSS.json |
113 | | - type: string |
114 | | - default: '' |
115 | | - required: False |
116 | | - benchmark_preset: |
117 | | - description: | |
118 | | - Name of benchmark preset to run. |
| 111 | + # Note: Currently, benchmark result filenames are in the format of |
| 112 | + # <benchmark_save_name>_<Device>_<Backend>_YYYYMMDD_HHMMSS.json |
| 113 | + # type: string |
| 114 | + # default: '' |
| 115 | + # required: False |
| 116 | + # benchmark_preset: |
| 117 | + # description: | |
| 118 | + # Name of benchmark preset to run. |
119 | 119 |
|
120 | | - See /devops/scripts/benchmarks/presets.py for all presets available. |
121 | | - type: string |
122 | | - default: 'Minimal' |
123 | | - required: False |
124 | | - # dry-run is passed only to compare.py (to not fail on regression), not to main.py (where such flag would omit all benchmark runs) |
125 | | - benchmark_dry_run: |
126 | | - description: | |
127 | | - Whether or not to fail the workflow upon a regression. |
128 | | - type: string |
129 | | - default: 'false' |
130 | | - required: False |
131 | | - benchmark_exit_on_failure: |
132 | | - description: | |
133 | | - Whether or not to fail the workflow upon a failure. |
134 | | - type: string |
135 | | - default: 'false' |
136 | | - required: False |
| 120 | + # See /devops/scripts/benchmarks/presets.py for all presets available. |
| 121 | + # type: string |
| 122 | + # default: 'Minimal' |
| 123 | + # required: False |
| 124 | + # # dry-run is passed only to compare.py (to not fail on regression), not to main.py (where such flag would omit all benchmark runs) |
| 125 | + # benchmark_dry_run: |
| 126 | + # description: | |
| 127 | + # Whether or not to fail the workflow upon a regression. |
| 128 | + # type: string |
| 129 | + # default: 'false' |
| 130 | + # required: False |
| 131 | + # benchmark_exit_on_failure: |
| 132 | + # description: | |
| 133 | + # Whether or not to fail the workflow upon a failure. |
| 134 | + # type: string |
| 135 | + # default: 'false' |
| 136 | + # required: False |
137 | 137 |
|
138 | | - in_workflow_call_mode: |
139 | | - description: | |
140 | | - Unlike `github.event_type` that captures the original trigger, we use |
141 | | - this to determine how this particular workflow has been called. This |
142 | | - isn't supposed to be specified by a caller and only relies on the |
143 | | - default value. |
144 | | - default: 'true' |
145 | | - required: False |
146 | | - type: string |
| 138 | + # in_workflow_call_mode: |
| 139 | + # description: | |
| 140 | + # Unlike `github.event_type` that captures the original trigger, we use |
| 141 | + # this to determine how this particular workflow has been called. This |
| 142 | + # isn't supposed to be specified by a caller and only relies on the |
| 143 | + # default value. |
| 144 | + # default: 'true' |
| 145 | + # required: False |
| 146 | + # type: string |
147 | 147 |
|
148 | 148 | workflow_dispatch: |
149 | 149 | inputs: |
|
0 commit comments