@@ -258,10 +258,11 @@ char InputByteSizeTest::input_data_string_
258258
259259TEST_F (InputByteSizeTest, ValidInputByteSize)
260260{
261+ const char * model_name = " savedmodel_zero_1_float32" ;
261262 // Create an inference request
262263 FAIL_TEST_IF_ERR (
263264 TRITONSERVER_InferenceRequestNew (
264- &irequest_, server_, " simple " , -1 /* model_version */ ),
265+ &irequest_, server_, model_name , -1 /* model_version */ ),
265266 " creating inference request" );
266267 FAIL_TEST_IF_ERR (
267268 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -312,10 +313,11 @@ TEST_F(InputByteSizeTest, ValidInputByteSize)
312313
313314TEST_F (InputByteSizeTest, InputByteSizeMismatch)
314315{
316+ const char * model_name = " savedmodel_zero_1_float32" ;
315317 // Create an inference request
316318 FAIL_TEST_IF_ERR (
317319 TRITONSERVER_InferenceRequestNew (
318- &irequest_, server_, " simple " , -1 /* model_version */ ),
320+ &irequest_, server_, model_name , -1 /* model_version */ ),
319321 " creating inference request" );
320322 FAIL_TEST_IF_ERR (
321323 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -353,8 +355,8 @@ TEST_F(InputByteSizeTest, InputByteSizeMismatch)
353355 FAIL_TEST_IF_SUCCESS (
354356 TRITONSERVER_ServerInferAsync (server_, irequest_, nullptr /* trace */ ),
355357 " expect error with inference request" ,
356- " input byte size mismatch for input 'INPUT0' for model 'simple'. "
357- " Expected 64, got 68" );
358+ " input byte size mismatch for input 'INPUT0' for model '" +
359+ std::string{model_name} + " '. Expected 64, got 68" );
358360
359361 // Need to manually delete request, otherwise server will not shut down.
360362 FAIL_TEST_IF_ERR (
@@ -364,10 +366,11 @@ TEST_F(InputByteSizeTest, InputByteSizeMismatch)
364366
365367TEST_F (InputByteSizeTest, ValidStringInputByteSize)
366368{
369+ const char * model_name = " savedmodel_zero_1_object" ;
367370 // Create an inference request
368371 FAIL_TEST_IF_ERR (
369372 TRITONSERVER_InferenceRequestNew (
370- &irequest_, server_, " simple_identity " , -1 /* model_version */ ),
373+ &irequest_, server_, model_name , -1 /* model_version */ ),
371374 " creating inference request" );
372375 FAIL_TEST_IF_ERR (
373376 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -424,10 +427,11 @@ TEST_F(InputByteSizeTest, ValidStringInputByteSize)
424427
425428TEST_F (InputByteSizeTest, StringCountMismatch)
426429{
430+ const char * model_name = " savedmodel_zero_1_object" ;
427431 // Create an inference request
428432 FAIL_TEST_IF_ERR (
429433 TRITONSERVER_InferenceRequestNew (
430- &irequest_, server_, " simple_identity " , -1 /* model_version */ ),
434+ &irequest_, server_, model_name , -1 /* model_version */ ),
431435 " creating inference request" );
432436 FAIL_TEST_IF_ERR (
433437 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -457,7 +461,8 @@ TEST_F(InputByteSizeTest, StringCountMismatch)
457461 FAIL_TEST_IF_SUCCESS (
458462 TRITONSERVER_ServerInferAsync (server_, irequest_, nullptr /* trace */ ),
459463 " expect error with inference request" ,
460- " expected 3 string elements for inference input 'INPUT0', got 2" );
464+ " expected 3 string elements for inference input 'INPUT0' for model '" +
465+ std::string{model_name} + " ', got 2" );
461466
462467 // Need to manually delete request, otherwise server will not shut down.
463468 FAIL_TEST_IF_ERR (
@@ -467,7 +472,8 @@ TEST_F(InputByteSizeTest, StringCountMismatch)
467472 // Create an inference request
468473 FAIL_TEST_IF_ERR (
469474 TRITONSERVER_InferenceRequestNew (
470- &irequest_, server_, " simple_identity" , -1 /* model_version */ ),
475+ &irequest_, server_, " savedmodel_zero_1_object" ,
476+ -1 /* model_version */ ),
471477 " creating inference request" );
472478 FAIL_TEST_IF_ERR (
473479 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -495,7 +501,9 @@ TEST_F(InputByteSizeTest, StringCountMismatch)
495501 FAIL_TEST_IF_SUCCESS (
496502 TRITONSERVER_ServerInferAsync (server_, irequest_, nullptr /* trace */ ),
497503 " expect error with inference request" ,
498- " expected 1 string elements for inference input 'INPUT0', got 2" );
504+ " unexpected number of string elements 2 for inference input 'INPUT0' for "
505+ " model '" +
506+ std::string{model_name} + " ', expecting 1" );
499507
500508 // Need to manually delete request, otherwise server will not shut down.
501509 FAIL_TEST_IF_ERR (
@@ -505,10 +513,11 @@ TEST_F(InputByteSizeTest, StringCountMismatch)
505513
506514TEST_F (InputByteSizeTest, StringSizeMisalign)
507515{
516+ const char * model_name = " savedmodel_zero_1_object" ;
508517 // Create an inference request
509518 FAIL_TEST_IF_ERR (
510519 TRITONSERVER_InferenceRequestNew (
511- &irequest_, server_, " simple_identity " , -1 /* model_version */ ),
520+ &irequest_, server_, model_name , -1 /* model_version */ ),
512521 " creating inference request" );
513522 FAIL_TEST_IF_ERR (
514523 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -542,9 +551,13 @@ TEST_F(InputByteSizeTest, StringSizeMisalign)
542551
543552 // Run inference
544553 FAIL_TEST_IF_SUCCESS (
545- TRITONSERVER_ServerInferAsync (server_, irequest_, nullptr /* trace
546- */ ), " expect error with inference request" ,
547- " element byte size indicator exceeds the end of the buffer" );
554+ TRITONSERVER_ServerInferAsync (server_, irequest_, nullptr /* trace*/ ),
555+ " expect error with inference request" ,
556+ " incomplete string length indicator for inference input 'INPUT0' for "
557+ " model '" +
558+ std::string{model_name} +
559+ " ', expecting 4 bytes but only 2 bytes available. Please make sure "
560+ " the string length indicator is in one buffer." );
548561
549562 // Need to manually delete request, otherwise server will not shut down.
550563 FAIL_TEST_IF_ERR (
@@ -573,7 +586,8 @@ TEST_F(InputByteSizeTest, StringCountMismatchGPU)
573586 // Create an inference request
574587 FAIL_TEST_IF_ERR (
575588 TRITONSERVER_InferenceRequestNew (
576- &irequest_, server_, " simple_identity" , -1 /* model_version */ ),
589+ &irequest_, server_, " savedmodel_zero_1_object" ,
590+ -1 /* model_version */ ),
577591 " creating inference request" );
578592 FAIL_TEST_IF_ERR (
579593 TRITONSERVER_InferenceRequestSetReleaseCallback (
@@ -629,7 +643,8 @@ TEST_F(InputByteSizeTest, StringCountMismatchGPU)
629643 // Create an inference request
630644 FAIL_TEST_IF_ERR (
631645 TRITONSERVER_InferenceRequestNew (
632- &irequest_, server_, " simple_identity" , -1 /* model_version */ ),
646+ &irequest_, server_, " savedmodel_zero_1_object" ,
647+ -1 /* model_version */ ),
633648 " creating inference request" );
634649 FAIL_TEST_IF_ERR (
635650 TRITONSERVER_InferenceRequestSetReleaseCallback (
0 commit comments