Skip to content

Commit e14bd58

Browse files
Add 1 more model and update script (#1527)
Signed-off-by: Tom Wildenhain <[email protected]>
1 parent 759b242 commit e14bd58

File tree

2 files changed

+42
-16
lines changed

2 files changed

+42
-16
lines changed

tests/cobalt_group_c_perf_testing_models.yaml

Lines changed: 28 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -90,11 +90,10 @@ cord-19-swivel-128d:
9090
converted_model: "D:/co_tfhubmodels_c/cord-19-swivel-128d/model.onnx"
9191
model_type: saved_model
9292
large_model: false
93-
run_tf_frozen: false
94-
use_custom_ops: true
95-
input_get: get_sentences
93+
input_get: get_ramp
9694
inputs:
9795
"sentences:0": [1]
96+
"256:0": [1, 512]
9897
outputs:
9998
- Identity:0
10099
rtol: 0.05
@@ -912,6 +911,7 @@ unet-industrial-class_10:
912911
- "UNet_v1/sigmoid:0"
913912
rtol: 0.05
914913
atol: 0.0005
914+
ptol: 1.0
915915
tag: ""
916916
signature_def: "default"
917917

@@ -1337,6 +1337,31 @@ on_device_vision-classifier-landmarks_classifier_africa_V1:
13371337
tag: ""
13381338
signature_def: "default"
13391339

1340+
tn_bert:
1341+
disabled: true
1342+
skip_conversion: true
1343+
model: "D:/co_tfhubmodels_c/tn_bert"
1344+
converted_model: "D:/co_tfhubmodels_c/tn_bert/model.onnx"
1345+
model_type: saved_model
1346+
large_model: false
1347+
input_get: get_small_rand_int32
1348+
inputs:
1349+
"input_word_ids:0": [1, 512]
1350+
"input_type_ids:0":
1351+
shape: [1, 512]
1352+
input_get: get_zeros_then_ones
1353+
"input_mask:0":
1354+
shape: [1, 512]
1355+
input_get: get_ones_int32
1356+
outputs:
1357+
- Identity:0
1358+
- Identity_1:0
1359+
- Identity_2:0
1360+
rtol: 0.05
1361+
atol: 0.0005
1362+
tag: "serve"
1363+
signature_def: "serving_default"
1364+
13401365
german-tacotron2:
13411366
disabled: true
13421367
skip_conversion: true

tests/run_pretrained_models.py

Lines changed: 14 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -419,19 +419,20 @@ def run_tflite():
419419
interpreter.invoke()
420420
result = [interpreter.get_tensor(output['index']) for output in output_details]
421421
return result
422-
tf_results = run_tflite()
423-
if self.perf:
424-
logger.info("Running TFLite perf")
425-
n = 0
426-
start = time.time()
427-
stop = start + PERF_TIME
428-
while time.time() < stop:
429-
for _ in range(PERF_STEP):
430-
_ = run_tflite()
431-
n += PERF_STEP
432-
self.tf_runtime = 1000 * (time.time() - start) / n
433-
logger.info("TFLite perf {:.2f}ms/inference, n={}".format(self.tf_runtime, n))
434-
logger.info("TFLite OK")
422+
if not self.skip_tensorflow:
423+
tf_results = run_tflite()
424+
if self.perf:
425+
logger.info("Running TFLite perf")
426+
n = 0
427+
start = time.time()
428+
stop = start + PERF_TIME
429+
while time.time() < stop:
430+
for _ in range(PERF_STEP):
431+
_ = run_tflite()
432+
n += PERF_STEP
433+
self.tf_runtime = 1000 * (time.time() - start) / n
434+
logger.info("TFLite perf {:.2f}ms/inference, n={}".format(self.tf_runtime, n))
435+
logger.info("TFLite OK")
435436

436437
if not self.run_tf_frozen:
437438
inputs = {}

0 commit comments

Comments
 (0)