31
31
def tearDownModule ():
32
32
output = json .dumps (result_data , indent = 4 )
33
33
if OUTPUT_FILE :
34
- with open (OUTPUT_FILE , "w" ) as opf :
34
+ with open (OUTPUT_FILE , "w" ) as opf : # noqa: PTH123
35
35
opf .write (output )
36
36
else :
37
- print (output )
37
+ print (output ) # noqa: T201
38
38
39
39
40
40
class Timer :
@@ -61,8 +61,9 @@ def tearDown(self):
61
61
name = self .__class__ .__name__ [4 :]
62
62
median = self .percentile (50 )
63
63
megabytes_per_sec = self .data_size / median / 1000000
64
- print (
65
- f"Completed { self .__class__ .__name__ } { megabytes_per_sec :.3f} MB/s, MEDIAN={ self .percentile (50 ):.3f} s, "
64
+ print ( # noqa: T201
65
+ f"Completed { self .__class__ .__name__ } { megabytes_per_sec :.3f} MB/s, "
66
+ f"MEDIAN={ self .percentile (50 ):.3f} s, "
66
67
f"total time={ duration :.3f} s, iterations={ len (self .results )} "
67
68
)
68
69
result_data .append (
@@ -119,7 +120,9 @@ def runTest(self):
119
120
with warnings .catch_warnings ():
120
121
warnings .simplefilter ("default" )
121
122
warnings .warn (
122
- f"{ self .__class__ .__name__ } timed out after { MAX_ITERATION_TIME } s, completed { i } /{ NUM_ITERATIONS } iterations."
123
+ f"{ self .__class__ .__name__ } timed out after { MAX_ITERATION_TIME } s, "
124
+ f"completed { i } /{ NUM_ITERATIONS } iterations." ,
125
+ stacklevel = 2 ,
123
126
)
124
127
125
128
break
@@ -132,7 +135,7 @@ class SmallFlatDocTest(PerformanceTest):
132
135
133
136
def setUp (self ):
134
137
super ().setUp ()
135
- with open (self .dataset ) as data :
138
+ with open (self .dataset ) as data : # noqa: PTH123
136
139
self .document = json .load (data )
137
140
138
141
self .data_size = len (encode (self .document )) * NUM_DOCS
@@ -204,14 +207,15 @@ def do_task(self):
204
207
def tearDown (self ):
205
208
super ().tearDown ()
206
209
SmallFlatModelFk .objects .all ().delete ()
210
+ ForeignKeyModel .objects .all ().delete ()
207
211
208
212
209
213
class LargeFlatDocTest (PerformanceTest ):
210
214
dataset = "large_doc.json"
211
215
212
216
def setUp (self ):
213
217
super ().setUp ()
214
- with open (self .dataset ) as data :
218
+ with open (self .dataset ) as data : # noqa: PTH123
215
219
self .document = json .load (data )
216
220
217
221
self .data_size = len (encode (self .document )) * NUM_DOCS
@@ -251,7 +255,7 @@ class LargeNestedDocTest(PerformanceTest):
251
255
252
256
def setUp (self ):
253
257
super ().setUp ()
254
- with open (self .dataset ) as data :
258
+ with open (self .dataset ) as data : # noqa: PTH123
255
259
self .document = json .load (data )
256
260
257
261
self .data_size = len (encode (self .document )) * NUM_DOCS
0 commit comments