File tree Expand file tree Collapse file tree 2 files changed +42
-16
lines changed Expand file tree Collapse file tree 2 files changed +42
-16
lines changed Original file line number Diff line number Diff line change @@ -90,11 +90,10 @@ cord-19-swivel-128d:
90
90
converted_model : " D:/co_tfhubmodels_c/cord-19-swivel-128d/model.onnx"
91
91
model_type : saved_model
92
92
large_model : false
93
- run_tf_frozen : false
94
- use_custom_ops : true
95
- input_get : get_sentences
93
+ input_get : get_ramp
96
94
inputs :
97
95
" sentences:0 " : [1]
96
+ " 256:0 " : [1, 512]
98
97
outputs :
99
98
- Identity:0
100
99
rtol : 0.05
@@ -912,6 +911,7 @@ unet-industrial-class_10:
912
911
- " UNet_v1/sigmoid:0"
913
912
rtol : 0.05
914
913
atol : 0.0005
914
+ ptol : 1.0
915
915
tag : " "
916
916
signature_def : " default"
917
917
@@ -1337,6 +1337,31 @@ on_device_vision-classifier-landmarks_classifier_africa_V1:
1337
1337
tag : " "
1338
1338
signature_def : " default"
1339
1339
1340
+ tn_bert :
1341
+ disabled : true
1342
+ skip_conversion : true
1343
+ model : " D:/co_tfhubmodels_c/tn_bert"
1344
+ converted_model : " D:/co_tfhubmodels_c/tn_bert/model.onnx"
1345
+ model_type : saved_model
1346
+ large_model : false
1347
+ input_get : get_small_rand_int32
1348
+ inputs :
1349
+ " input_word_ids:0 " : [1, 512]
1350
+ " input_type_ids:0 " :
1351
+ shape : [1, 512]
1352
+ input_get : get_zeros_then_ones
1353
+ " input_mask:0 " :
1354
+ shape : [1, 512]
1355
+ input_get : get_ones_int32
1356
+ outputs :
1357
+ - Identity:0
1358
+ - Identity_1:0
1359
+ - Identity_2:0
1360
+ rtol : 0.05
1361
+ atol : 0.0005
1362
+ tag : " serve"
1363
+ signature_def : " serving_default"
1364
+
1340
1365
german-tacotron2 :
1341
1366
disabled : true
1342
1367
skip_conversion : true
Original file line number Diff line number Diff line change @@ -419,19 +419,20 @@ def run_tflite():
419
419
interpreter .invoke ()
420
420
result = [interpreter .get_tensor (output ['index' ]) for output in output_details ]
421
421
return result
422
- tf_results = run_tflite ()
423
- if self .perf :
424
- logger .info ("Running TFLite perf" )
425
- n = 0
426
- start = time .time ()
427
- stop = start + PERF_TIME
428
- while time .time () < stop :
429
- for _ in range (PERF_STEP ):
430
- _ = run_tflite ()
431
- n += PERF_STEP
432
- self .tf_runtime = 1000 * (time .time () - start ) / n
433
- logger .info ("TFLite perf {:.2f}ms/inference, n={}" .format (self .tf_runtime , n ))
434
- logger .info ("TFLite OK" )
422
+ if not self .skip_tensorflow :
423
+ tf_results = run_tflite ()
424
+ if self .perf :
425
+ logger .info ("Running TFLite perf" )
426
+ n = 0
427
+ start = time .time ()
428
+ stop = start + PERF_TIME
429
+ while time .time () < stop :
430
+ for _ in range (PERF_STEP ):
431
+ _ = run_tflite ()
432
+ n += PERF_STEP
433
+ self .tf_runtime = 1000 * (time .time () - start ) / n
434
+ logger .info ("TFLite perf {:.2f}ms/inference, n={}" .format (self .tf_runtime , n ))
435
+ logger .info ("TFLite OK" )
435
436
436
437
if not self .run_tf_frozen :
437
438
inputs = {}
You can’t perform that action at this time.
0 commit comments