@@ -204,12 +204,12 @@ def test_opt_levels(args):
204
204
return 0
205
205
206
206
207
- def measure (driver , tests , i ):
207
+ def measure (driver , tests , i , min_num_samples ):
208
208
"""Log and measure samples of the tests with the given driver.
209
209
210
210
Collect increasing number of samples, depending on the iteration.
211
211
"""
212
- num_samples = min (i + 3 , 10 )
212
+ num_samples = max ( min (i + 3 , 10 ), min_num_samples )
213
213
msg = " Iteration {0} for {1}: num samples = {2}, " .format (
214
214
i , driver .args .tests , num_samples
215
215
)
@@ -246,7 +246,7 @@ def test_performance(
246
246
optimization = opt_level ))
247
247
for dir in [old_dir , new_dir ]
248
248
]
249
- results = [measure (driver , driver .tests , i ) for driver in [old , new ]]
249
+ results = [measure (driver , driver .tests , i , num_samples ) for driver in [old , new ]]
250
250
tests = TestComparator (results [0 ], results [1 ], threshold )
251
251
changed = tests .decreased + tests .increased
252
252
@@ -255,7 +255,7 @@ def test_performance(
255
255
if VERBOSE :
256
256
log (" test again: " + str ([test .name for test in changed ]))
257
257
results = [
258
- merge (the_results , measure (driver , [test .name for test in changed ], i ))
258
+ merge (the_results , measure (driver , [test .name for test in changed ], i , num_samples ))
259
259
for the_results , driver in zip (results , [old , new ])
260
260
]
261
261
tests = TestComparator (results [0 ], results [1 ], threshold )
0 commit comments