@@ -105,7 +105,7 @@ size_t dispatchNVFP4xNVFP4GemmClusterShapeSm100(T* D, void const* A, void const*
105105 break ;
106106 default :
107107 throw std::runtime_error (
108- " [TensorRT- LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
108+ " [TensorRT LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
109109 break ;
110110 }
111111}
@@ -146,15 +146,15 @@ size_t dispatchNVFP4xNVFP4GemmCTAShapeSm100(T* D, void const* A, void const* B,
146146 occupancy);
147147 break ;
148148 case tkc::CutlassTileConfigSM100::Undefined:
149- throw std::runtime_error (" [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config undefined." );
149+ throw std::runtime_error (" [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config undefined." );
150150 break ;
151151 case tkc::CutlassTileConfigSM100::ChooseWithHeuristic:
152152 throw std::runtime_error (
153- " [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config should have already been set by "
153+ " [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config should have already been set by "
154154 " heuristic." );
155155 break ;
156156 default :
157- throw std::runtime_error (" [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
157+ throw std::runtime_error (" [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
158158 break ;
159159 }
160160}
@@ -177,7 +177,7 @@ size_t dispatchNVFP4xNVFP4GemmClusterShapeSm120(T* D, void const* A, void const*
177177 break ;
178178 default :
179179 throw std::runtime_error (
180- " [TensorRT- LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
180+ " [TensorRT LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
181181 break ;
182182 }
183183}
@@ -205,16 +205,16 @@ size_t dispatchNVFP4xNVFP4GemmCTAShapeSm120(T* D, void const* A, void const* B,
205205 occupancy);
206206 break ;
207207 case tkc::CutlassTileConfigSM120::Undefined:
208- throw std::runtime_error (" [TensorRT- LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config undefined." );
208+ throw std::runtime_error (" [TensorRT LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config undefined." );
209209 break ;
210210 case tkc::CutlassTileConfigSM120::ChooseWithHeuristic:
211211 throw std::runtime_error (
212- " [TensorRT- LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config should have already been set by "
212+ " [TensorRT LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Gemm config should have already been set by "
213213 " heuristic." );
214214 break ;
215215 default :
216216 throw std::runtime_error (
217- " [TensorRT- LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
217+ " [TensorRT LLM Error][FP4][sm120][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
218218 break ;
219219 }
220220}
@@ -257,7 +257,7 @@ size_t dispatchMXFP8xMXFP4GemmClusterShapeSm100(T* D, void const* A, void const*
257257 break ;
258258 default :
259259 throw std::runtime_error (
260- " [TensorRT- LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
260+ " [TensorRT LLM Error][FP4][dispatch_gemm_cluster_shape] Config is invalid for FP4 GEMM." );
261261 break ;
262262 }
263263}
@@ -293,15 +293,15 @@ size_t dispatchMXFP8xMXFP4GemmCTAShapeSm100(T* D, void const* A, void const* B,
293293 occupancy);
294294 break ;
295295 case tkc::CutlassTileConfigSM100::Undefined:
296- throw std::runtime_error (" [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config undefined." );
296+ throw std::runtime_error (" [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config undefined." );
297297 break ;
298298 case tkc::CutlassTileConfigSM100::ChooseWithHeuristic:
299299 throw std::runtime_error (
300- " [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config should have already been set by "
300+ " [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Gemm config should have already been set by "
301301 " heuristic." );
302302 break ;
303303 default :
304- throw std::runtime_error (" [TensorRT- LLM Error][FP4][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
304+ throw std::runtime_error (" [TensorRT LLM Error][FP4][dispatch_gemm_cta_shape] Config is invalid for FP4 GEMM." );
305305 break ;
306306 }
307307}
@@ -338,7 +338,7 @@ size_t CutlassFp4GemmRunner<T, fp4GemmType>::dispatchToArch(T* D, void const* A,
338338 else
339339 {
340340 throw std::runtime_error (
341- " [TensorRT- LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
341+ " [TensorRT LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
342342 }
343343 }
344344 else if constexpr (fp4GemmType == FP4GemmType::W4A4_NVFP4_NVFP4)
@@ -356,13 +356,13 @@ size_t CutlassFp4GemmRunner<T, fp4GemmType>::dispatchToArch(T* D, void const* A,
356356 else
357357 {
358358 throw std::runtime_error (
359- " [TensorRT- LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
359+ " [TensorRT LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] Arch unsupported for CUTLASS FP4 GEMM" );
360360 }
361361 }
362362 else
363363 {
364364 throw std::runtime_error (
365- " [TensorRT- LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] FP4 Gemm type unsupported for CUTLASS FP4 GEMM" );
365+ " [TensorRT LLM Error][CutlassFp4GemmRunner][GEMM Dispatch] FP4 Gemm type unsupported for CUTLASS FP4 GEMM" );
366366 }
367367}
368368
0 commit comments