@@ -149,8 +149,7 @@ async def write_s3_obj(app, obj_id, bucket=None):
149149 bucket = domain_bucket
150150
151151 if obj_id in pending_s3_write :
152- msg = f"write_s3_key - not expected for key { obj_id } to be in "
153- msg += "pending_s3_write map"
152+ msg = f"write_s3_key - not expected for key { obj_id } to be in pending_s3_write map"
154153 log .error (msg )
155154 raise KeyError (msg )
156155
@@ -172,12 +171,10 @@ async def write_s3_obj(app, obj_id, bucket=None):
172171 # timestamp is first element of two-tuple
173172 last_update_time = dirty_ids [obj_id ][0 ]
174173 else :
175- msg = f"write_s3_obj - { obj_id } not in dirty_ids, "
176- msg += "assuming flush write"
174+ msg = f"write_s3_obj - { obj_id } not in dirty_ids, assuming flush write"
177175 log .debug (msg )
178176 if last_update_time > now :
179- msg = f"last_update time { last_update_time } is in the future for "
180- msg += f"obj_id: { obj_id } "
177+ msg = f"last_update time { last_update_time } is in the future for obj_id: { obj_id } "
181178 log .error (msg )
182179 raise ValueError (msg )
183180
@@ -198,8 +195,7 @@ async def write_s3_obj(app, obj_id, bucket=None):
198195 dset_id = getDatasetId (obj_id )
199196 if dset_id in filter_map :
200197 filter_ops = filter_map [dset_id ]
201- msg = f"write_s3_obj: got filter_op: { filter_ops } "
202- msg += f"for dset: { dset_id } "
198+ msg = f"write_s3_obj: got filter_op: { filter_ops } for dset: { dset_id } "
203199 log .debug (msg )
204200 else :
205201 filter_ops = None
@@ -237,13 +233,11 @@ async def write_s3_obj(app, obj_id, bucket=None):
237233 # meta data update
238234 # check for object in meta cache
239235 if obj_id not in meta_cache :
240- msg = f"write_s3_obj: expected to find obj_id: { obj_id } "
241- msg += "in meta cache"
236+ msg = f"write_s3_obj: expected to find obj_id: { obj_id } in meta cache"
242237 log .error (msg )
243238 raise KeyError (f"{ obj_id } not found in meta cache" )
244239 if not meta_cache .isDirty (obj_id ):
245- msg = f"write_s3_obj: expected meta cache obj { obj_id } "
246- msg == "to be dirty"
240+ msg = f"write_s3_obj: expected meta cache obj { obj_id } to be dirty"
247241 log .error (msg )
248242 raise ValueError ("bad dirty state for obj" )
249243 obj_json = meta_cache [obj_id ]
@@ -264,8 +258,7 @@ async def write_s3_obj(app, obj_id, bucket=None):
264258 else :
265259 timestamp = 0
266260 if timestamp > last_update_time :
267- msg = f"write_s3_obj: { obj_id } got updated while s3 "
268- msg += "write was in progress"
261+ msg = f"write_s3_obj: { obj_id } got updated while s3 write was in progress"
269262 log .info (msg )
270263 else :
271264 log .debug (f"write_s3obj: clear dirty for { obj_id } " )
@@ -279,11 +272,10 @@ async def write_s3_obj(app, obj_id, bucket=None):
279272
280273 finally :
281274 # clear pending_s3_write item
282- log .debug (f"write_s3_obj finally block, success={ success } " )
275+ log .debug (f"write_s3_obj { obj_id } finally block, success={ success } " )
283276 if obj_id in pending_s3_write :
284277 if pending_s3_write [obj_id ] != now :
285- msg = "pending_s3_write timestamp got updated unexpectedly "
286- msg += f"for { obj_id } "
278+ msg = f"pending_s3_write timestamp got updated unexpectedly for { obj_id } "
287279 log .error (msg )
288280 del pending_s3_write [obj_id ]
289281 # clear task
@@ -1259,10 +1251,9 @@ def callback(future):
12591251
12601252 if obj_id in pending_s3_write :
12611253 pending_time = s3sync_start - pending_s3_write [obj_id ]
1262- msg = f"s3sync - key { obj_id } has been pending for "
1263- msg += f"{ pending_time :.3f} "
1254+ msg = f"s3sync - key { obj_id } has been pending for { pending_time :.3f} "
12641255 log .debug (msg )
1265- if s3sync_start - pending_s3_write [ obj_id ] > s3_sync_task_timeout :
1256+ if pending_time > s3_sync_task_timeout :
12661257 msg = f"s3sync - obj { obj_id } has been in pending_s3_write "
12671258 msg += f"for { pending_time :.3f} seconds, restarting"
12681259 log .warn (msg )
0 commit comments