|
214 | 214 | - arg_meta: null
|
215 | 215 | kernel_name: impl::reference::quantized_linear_out
|
216 | 216 |
|
| 217 | +- func: cadence::quantized_linear.per_tensor_out(Tensor src, Tensor weight, Tensor bias, SymInt src_zero_point, SymInt weight_zero_point, SymInt out_multiplier, SymInt out_shift, SymInt out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!) |
| 218 | + kernels: |
| 219 | + - arg_meta: null |
| 220 | + kernel_name: impl::reference::quantized_linear_per_tensor_out |
| 221 | + |
| 222 | +- func: cadence::quantized_linear_asym8sxasym8s_asym8s.per_tensor_out(Tensor src, Tensor weight, Tensor bias, int src_zero_point, int weight_zero_point, int out_multiplier, int out_shift, int out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!) |
| 223 | + kernels: |
| 224 | + - arg_meta: null |
| 225 | + kernel_name: impl::reference::quantized_linear_asym8sxasym8s_asym8s_per_tensor_out |
| 226 | + |
| 227 | +- func: cadence::quantized_linear_asym8uxasym8u_asym8u.per_tensor_out(Tensor src, Tensor weight, Tensor bias, int src_zero_point, int weight_zero_point, int out_multiplier, int out_shift, int out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!) |
| 228 | + kernels: |
| 229 | + - arg_meta: null |
| 230 | + kernel_name: impl::reference::quantized_linear_asym8uxasym8u_asym8u_per_tensor_out |
| 231 | + |
217 | 232 | - func: cadence::quantized_relu.out(Tensor X, Tensor X_zero_point, int out_zero_point, Tensor out_multiplier, Tensor out_shift, *, Tensor(a!) out) -> Tensor(a!)
|
218 | 233 | kernels:
|
219 | 234 | - arg_meta: null
|
|
249 | 264 | - arg_meta: null
|
250 | 265 | kernel_name: impl::reference::quantized_matmul_asym8uxasym8u_asym8u_out
|
251 | 266 |
|
252 |
| -- func: cadence::quantized_linear.per_tensor_out(Tensor src, Tensor weight, Tensor bias, SymInt src_zero_point, SymInt weight_zero_point, SymInt out_multiplier, SymInt out_shift, SymInt out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!) |
| 267 | +- func: cadence::im2row.out(Tensor input, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, Tensor in_zero_point, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
253 | 268 | kernels:
|
254 | 269 | - arg_meta: null
|
255 |
| - kernel_name: impl::reference::quantized_linear_per_tensor_out |
| 270 | + kernel_name: impl::reference::im2row_out |
256 | 271 |
|
257 |
| -- func: cadence::quantized_linear_asym8sxasym8s_asym8s.per_tensor_out(Tensor src, Tensor weight, Tensor bias, int src_zero_point, int weight_zero_point, int out_multiplier, int out_shift, int out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!) |
| 272 | +- func: cadence::im2row.per_tensor_out(Tensor input, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, int in_zero_point, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
258 | 273 | kernels:
|
259 | 274 | - arg_meta: null
|
260 |
| - kernel_name: impl::reference::quantized_linear_asym8sxasym8s_asym8s_per_tensor_out |
| 275 | + kernel_name: impl::reference::im2row_per_tensor_out |
261 | 276 |
|
262 |
| -- func: cadence::quantized_linear_asym8uxasym8u_asym8u.per_tensor_out(Tensor src, Tensor weight, Tensor bias, int src_zero_point, int weight_zero_point, int out_multiplier, int out_shift, int out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!) |
| 277 | +- func: cadence::quantized_conv_nchw.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
263 | 278 | kernels:
|
264 | 279 | - arg_meta: null
|
265 |
| - kernel_name: impl::reference::quantized_linear_asym8uxasym8u_asym8u_per_tensor_out |
| 280 | + kernel_name: impl::reference::quantized_conv_nchw_per_tensor_out |
266 | 281 |
|
267 |
| -- func: cadence::im2row.out(Tensor input, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, Tensor in_zero_point, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
| 282 | +- func: cadence::quantized_conv_nhwc.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
268 | 283 | kernels:
|
269 | 284 | - arg_meta: null
|
270 |
| - kernel_name: impl::reference::im2row_out |
| 285 | + kernel_name: impl::reference::quantized_conv_nhwc_per_tensor_out |
271 | 286 |
|
272 |
| -- func: cadence::im2row.per_tensor_out(Tensor input, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, int in_zero_point, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
| 287 | +- func: cadence::quantized_conv_nchw_asym8sxsym8s_asym8s.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, *, Tensor(a!) out) -> Tensor(a!) |
273 | 288 | kernels:
|
274 | 289 | - arg_meta: null
|
275 |
| - kernel_name: impl::reference::im2row_per_tensor_out |
| 290 | + kernel_name: impl::reference::quantized_conv_nchw_asym8sxsym8s_asym8s_per_tensor_out |
276 | 291 |
|
277 |
| -- func: cadence::quantized_conv_nchw.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
| 292 | +- func: cadence::quantized_conv_nchw_asym8uxsym8u_asym8u.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, *, Tensor(a!) out) -> Tensor(a!) |
278 | 293 | kernels:
|
279 | 294 | - arg_meta: null
|
280 |
| - kernel_name: impl::reference::quantized_conv_nchw_per_tensor_out |
| 295 | + kernel_name: impl::reference::quantized_conv_nchw_asym8uxsym8u_asym8u_per_tensor_out |
281 | 296 |
|
282 |
| -- func: cadence::quantized_conv_nhwc.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, bool channel_last=False, *, Tensor(a!) out) -> Tensor(a!) |
| 297 | +- func: cadence::quantized_conv_nhwc_asym8sxsym8s_asym8s.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, *, Tensor(a!) out) -> Tensor(a!) |
283 | 298 | kernels:
|
284 | 299 | - arg_meta: null
|
285 |
| - kernel_name: impl::reference::quantized_conv_nhwc_per_tensor_out |
| 300 | + kernel_name: impl::reference::quantized_conv_nhwc_asym8sxsym8s_asym8s_per_tensor_out |
| 301 | + |
| 302 | +- func: cadence::quantized_conv_nhwc_asym8uxsym8u_asym8u.per_tensor_out(Tensor input, Tensor weight, Tensor bias, int[] stride, SymInt[] padding, int[] dilation, int groups, int input_zero_point, int weight_zero_point, float bias_scale, float out_scale, int out_zero_point, int out_multiplier, int out_shift, *, Tensor(a!) out) -> Tensor(a!) |
| 303 | + kernels: |
| 304 | + - arg_meta: null |
| 305 | + kernel_name: impl::reference::quantized_conv_nhwc_asym8uxsym8u_asym8u_per_tensor_out |
286 | 306 |
|
287 | 307 | - func: cadence::quantized_fully_connected.out(Tensor src, Tensor weight, Tensor bias, int src_zero_point, Tensor weight_zero_point, Tensor out_multiplier, Tensor out_shift, int out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!)
|
288 | 308 | kernels:
|
|
0 commit comments