@@ -141,23 +141,9 @@ InferResponseComplete(
141
141
output_tensors.push_back (pb_tensor);
142
142
}
143
143
}
144
- }
145
- catch (const PythonBackendException& pb_exception) {
146
- if (response != nullptr ) {
147
- LOG_IF_ERROR (
148
- TRITONSERVER_InferenceResponseDelete (response),
149
- " Failed to delete inference response." );
150
144
151
- response = nullptr ;
152
- }
153
- pb_error = std::make_shared<PbError>(pb_exception.what ());
154
- output_tensors.clear ();
155
- }
156
-
157
- try {
158
145
triton::common::TritonJson::Value parameters_json (
159
146
triton::common::TritonJson::ValueType::OBJECT);
160
- std::cerr << " debug ziqif: response = " << response << std::endl;
161
147
uint32_t parameter_count;
162
148
THROW_IF_TRITON_ERROR (TRITONSERVER_InferenceResponseParameterCount (
163
149
response, ¶meter_count));
@@ -185,10 +171,6 @@ InferResponseComplete(
185
171
name + " '." ));
186
172
}
187
173
}
188
-
189
- triton::common::TritonJson::WriteBuffer buffer;
190
- THROW_IF_TRITON_ERROR (parameters_json.Write (&buffer));
191
- parameters_string = buffer.Contents ();
192
174
}
193
175
catch (const PythonBackendException& pb_exception) {
194
176
if (response != nullptr ) {
@@ -199,6 +181,7 @@ InferResponseComplete(
199
181
response = nullptr ;
200
182
}
201
183
pb_error = std::make_shared<PbError>(pb_exception.what ());
184
+ output_tensors.clear ();
202
185
}
203
186
204
187
if (!infer_payload->IsDecoupled ()) {
0 commit comments