Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion backends/cadence/fusion_g3/operators/op_clamp.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -21,13 +21,13 @@
#include <executorch/kernels/portable/cpu/util/math_util.h>
#include <executorch/runtime/kernel/kernel_includes.h>

using ::executorch::aten::optional;
using ::executorch::aten::Scalar;
using ::executorch::aten::ScalarType;
using ::executorch::aten::Tensor;
using ::executorch::runtime::canCast;
using ::executorch::runtime::Error;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;

namespace cadence {
namespace impl {
Expand Down
20 changes: 10 additions & 10 deletions backends/cadence/fusion_g3/operators/op_dequantize.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ using ::executorch::runtime::Error;
using ::executorch::runtime::KernelRuntimeContext;

template <typename T>
using optional = ::executorch::aten::optional<T>;
using optional = std::optional<T>;
/* ScalarType in Executorch do not have support for below data types.
* So, creating a placeholder for these data types. Once, ScalarTypes is
* updated to have support for below data types, these can be removed and
Expand All @@ -51,7 +51,7 @@ void check_dequantize_per_tensor_args(
int64_t quant_min,
int64_t quant_max,
ScalarType dtype,
::executorch::aten::optional<ScalarType>& out_dtype,
std::optional<ScalarType>& out_dtype,
Tensor& out) {
ET_CHECK_MSG(
input.scalar_type() == ScalarType::Byte ||
Expand Down Expand Up @@ -93,7 +93,7 @@ Tensor& dequantize_impl(
float* scale_data,
int* zero_point_data,
int* axis,
::executorch::aten::optional<ScalarType> out_dtype) {
std::optional<ScalarType> out_dtype) {
const ::executorch::aten::ArrayRef<Tensor::SizesType> input_size =
input.sizes();

Expand Down Expand Up @@ -260,8 +260,8 @@ Tensor& dequantize_impl(
}
}

::executorch::aten::optional<::executorch::aten::ArrayRef<int64_t>>
optional_dim_list{::executorch::aten::ArrayRef<int64_t>{
std::optional<::executorch::aten::ArrayRef<int64_t>> optional_dim_list{
::executorch::aten::ArrayRef<int64_t>{
dims, size_t(input.dim() - 1)}};

// Actual dequantization logic
Expand Down Expand Up @@ -466,8 +466,8 @@ Tensor& dequantize_impl(
}
}

::executorch::aten::optional<::executorch::aten::ArrayRef<int64_t>>
optional_dim_list{::executorch::aten::ArrayRef<int64_t>{
std::optional<::executorch::aten::ArrayRef<int64_t>> optional_dim_list{
::executorch::aten::ArrayRef<int64_t>{
dims, size_t(input.dim() - 1)}};

// Actual dequantization logic
Expand Down Expand Up @@ -600,7 +600,7 @@ Tensor& dequantize_per_tensor_tensor_args_out(
int64_t quant_min,
int64_t quant_max,
ScalarType dtype,
::executorch::aten::optional<ScalarType> out_dtype,
std::optional<ScalarType> out_dtype,
Tensor& out) {
#ifdef OP_ARG_CHECK
ET_CHECK_MSG(
Expand Down Expand Up @@ -639,12 +639,12 @@ Tensor& dequantize_per_channel_out(
KernelRuntimeContext& context,
const Tensor& input,
const Tensor& scale,
const ::executorch::aten::optional<Tensor>& opt_zero_points,
const std::optional<Tensor>& opt_zero_points,
int64_t axis,
int64_t quant_min,
int64_t quant_max,
ScalarType dtype,
::executorch::aten::optional<ScalarType> out_dtype,
std::optional<ScalarType> out_dtype,
Tensor& out) {
if (axis < 0) {
axis += executorch::runtime::nonzero_dim(input);
Expand Down
6 changes: 3 additions & 3 deletions backends/cadence/fusion_g3/operators/op_div.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -19,14 +19,14 @@
#include <executorch/runtime/kernel/kernel_includes.h>
#include <executorch/runtime/platform/assert.h>

using ::executorch::aten::optional;
using ::executorch::aten::Scalar;
using ::executorch::aten::ScalarType;
using ::executorch::aten::string_view;
using ::executorch::aten::Tensor;
using ::executorch::runtime::canCast;
using ::executorch::runtime::Error;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;
using std::string_view;

namespace cadence {
namespace impl {
Expand Down Expand Up @@ -686,4 +686,4 @@ Tensor& div_scalar_mode_out(
} // namespace native
} // namespace G3
} // namespace impl
} // namespace cadence
} // namespace cadence
2 changes: 1 addition & 1 deletion backends/cadence/fusion_g3/operators/op_mean.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -17,11 +17,11 @@
#include <executorch/runtime/platform/assert.h>

using ::executorch::aten::ArrayRef;
using ::executorch::aten::optional;
using ::executorch::aten::ScalarType;
using ::executorch::aten::Tensor;
using ::executorch::runtime::Error;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;

namespace cadence {
namespace impl {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,11 +19,11 @@
#include <executorch/runtime/kernel/kernel_includes.h>

using ::executorch::aten::IntArrayRef;
using ::executorch::aten::optional;
using ::executorch::aten::ScalarType;
using ::executorch::aten::Tensor;
using ::executorch::runtime::Error;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;

namespace cadence {
namespace impl {
Expand Down
8 changes: 4 additions & 4 deletions backends/cadence/fusion_g3/operators/op_quantize.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -329,8 +329,8 @@ Tensor& quantize_impl(
}
}

::executorch::aten::optional<::executorch::aten::ArrayRef<int64_t>>
optional_dim_list{::executorch::aten::ArrayRef<int64_t>{
std::optional<::executorch::aten::ArrayRef<int64_t>> optional_dim_list{
::executorch::aten::ArrayRef<int64_t>{
dims, size_t(input.dim() - 1)}};

// Actual quantization logic
Expand Down Expand Up @@ -534,8 +534,8 @@ Tensor& quantize_impl(
}
}

::executorch::aten::optional<::executorch::aten::ArrayRef<int64_t>>
optional_dim_list{::executorch::aten::ArrayRef<int64_t>{
std::optional<::executorch::aten::ArrayRef<int64_t>> optional_dim_list{
::executorch::aten::ArrayRef<int64_t>{
dims, size_t(input.dim() - 1)}};

// Actual quantization logic
Expand Down
4 changes: 2 additions & 2 deletions backends/cadence/fusion_g3/operators/op_slice_copy.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -37,8 +37,8 @@ Tensor& slice_copy_Tensor_out(
KernelRuntimeContext& ctx,
const Tensor& in,
int64_t dim,
::executorch::aten::optional<int64_t> start_val,
::executorch::aten::optional<int64_t> end_val,
std::optional<int64_t> start_val,
std::optional<int64_t> end_val,
int64_t step,
Tensor& out) {
(void)ctx;
Expand Down
32 changes: 15 additions & 17 deletions backends/cadence/fusion_g3/operators/operators.h
Original file line number Diff line number Diff line change
Expand Up @@ -47,13 +47,12 @@ ::executorch::aten::Tensor& dequantize_per_channel_out(
::executorch::runtime::KernelRuntimeContext& context,
const ::executorch::aten::Tensor& input,
const ::executorch::aten::Tensor& scale,
const ::executorch::aten::optional<::executorch::aten::Tensor>&
opt_zero_points,
const std::optional<::executorch::aten::Tensor>& opt_zero_points,
int64_t axis,
int64_t quant_min,
int64_t quant_max,
::executorch::aten::ScalarType dtype,
::executorch::aten::optional<::executorch::aten::ScalarType> out_dtype,
std::optional<::executorch::aten::ScalarType> out_dtype,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& dequantize_per_tensor_out(
Expand All @@ -64,7 +63,7 @@ ::executorch::aten::Tensor& dequantize_per_tensor_out(
int64_t quant_min,
int64_t quant_max,
::executorch::aten::ScalarType dtype,
::executorch::aten::optional<::executorch::aten::ScalarType> out_dtype,
std::optional<::executorch::aten::ScalarType> out_dtype,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& div_out(
Expand All @@ -77,7 +76,7 @@ ::executorch::aten::Tensor& div_out_mode(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& a,
const ::executorch::aten::Tensor& b,
::executorch::aten::optional<::executorch::aten::string_view> mode,
std::optional<std::string_view> mode,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& div_scalar_out(
Expand All @@ -90,7 +89,7 @@ ::executorch::aten::Tensor& div_scalar_mode_out(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& a,
const ::executorch::aten::Scalar& b,
::executorch::aten::optional<::executorch::aten::string_view> mode,
std::optional<std::string_view> mode,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& exp_out(
Expand All @@ -101,10 +100,9 @@ ::executorch::aten::Tensor& exp_out(
::executorch::aten::Tensor& mean_dim_out(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& in,
::executorch::aten::optional<::executorch::aten::ArrayRef<int64_t>>
dim_list,
std::optional<::executorch::aten::ArrayRef<int64_t>> dim_list,
bool keepdim,
::executorch::aten::optional<::executorch::aten::ScalarType> dtype,
std::optional<::executorch::aten::ScalarType> dtype,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& mul_out(
Expand All @@ -127,8 +125,8 @@ native_layer_norm_out(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& input,
::executorch::aten::IntArrayRef normalized_shape,
const ::executorch::aten::optional<::executorch::aten::Tensor>& weight,
const ::executorch::aten::optional<::executorch::aten::Tensor>& bias,
const std::optional<::executorch::aten::Tensor>& weight,
const std::optional<::executorch::aten::Tensor>& bias,
double eps,
::executorch::aten::Tensor& out,
::executorch::aten::Tensor& mean_out,
Expand Down Expand Up @@ -165,8 +163,8 @@ ::executorch::aten::Tensor& slice_copy_Tensor_out(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& in,
int64_t dim,
::executorch::aten::optional<int64_t> start_val,
::executorch::aten::optional<int64_t> end_val,
std::optional<int64_t> start_val,
std::optional<int64_t> end_val,
int64_t step,
::executorch::aten::Tensor& out);

Expand Down Expand Up @@ -226,15 +224,15 @@ ::executorch::aten::Tensor& where_out(
::executorch::aten::Tensor& clamp_out(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& in,
const ::executorch::aten::optional<::executorch::aten::Scalar>& min_opt,
const ::executorch::aten::optional<::executorch::aten::Scalar>& max_opt,
const std::optional<::executorch::aten::Scalar>& min_opt,
const std::optional<::executorch::aten::Scalar>& max_opt,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& clamp_tensor_out(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& in,
const ::executorch::aten::optional<::executorch::aten::Tensor>& min_opt,
const ::executorch::aten::optional<::executorch::aten::Tensor>& max_opt,
const std::optional<::executorch::aten::Tensor>& min_opt,
const std::optional<::executorch::aten::Tensor>& max_opt,
::executorch::aten::Tensor& out);

::executorch::aten::Tensor& transpose_copy_int_out(
Expand Down
8 changes: 4 additions & 4 deletions backends/cadence/hifi/operators/op_clamp.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -51,8 +51,8 @@ namespace native {
Tensor& clamp_Tensor_out(
RuntimeContext& ctx,
const Tensor& in,
const executorch::aten::optional<Tensor>& min_opt,
const executorch::aten::optional<Tensor>& max_opt,
const std::optional<Tensor>& min_opt,
const std::optional<Tensor>& max_opt,
Tensor& out) {
(void)ctx;

Expand Down Expand Up @@ -325,8 +325,8 @@ Tensor& clamp_Tensor_out(
Tensor& clamp_tensor_out(
RuntimeContext& ctx,
const Tensor& in,
const executorch::aten::optional<Tensor>& min_opt,
const executorch::aten::optional<Tensor>& max_opt,
const std::optional<Tensor>& min_opt,
const std::optional<Tensor>& max_opt,
Tensor& out) {
return clamp_Tensor_out(ctx, in, min_opt, max_opt, out);
}
Expand Down
2 changes: 1 addition & 1 deletion backends/cadence/hifi/operators/op_div.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -178,7 +178,7 @@ Tensor& div_out_mode(
RuntimeContext& ctx,
const Tensor& a,
const Tensor& b,
executorch::aten::optional<executorch::aten::string_view> mode,
std::optional<std::string_view> mode,
Tensor& out) {
ET_KERNEL_CHECK(
ctx,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,12 +19,12 @@ namespace native {

using ::executorch::aten::ArrayRef;
using ::executorch::aten::IntArrayRef;
using ::executorch::aten::optional;
using ::executorch::aten::Scalar;
using ::executorch::aten::ScalarType;
using ::executorch::aten::SizesType;
using ::executorch::aten::Tensor;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;

void inline _quantized_fully_connected_asym8u(
const Tensor& in,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,11 +20,11 @@ namespace impl {
namespace HiFi {
namespace native {

using ::executorch::aten::optional;
using ::executorch::aten::ScalarType;
using ::executorch::aten::Tensor;
using ::executorch::runtime::getLeadingDims;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;

// The nnlib kernel to compute quantized linear via matmul.

Expand Down
2 changes: 1 addition & 1 deletion backends/cadence/hifi/operators/op_softmax.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ Tensor& _softmax_out(
// Adjust for negative dim
dim = dim < 0 ? dim + executorch::runtime::nonzero_dim(in) : dim;

const executorch::aten::optional<int64_t>& dim_t = dim;
const std::optional<int64_t>& dim_t = dim;
const size_t d = ET_NORMALIZE_IX(dim_t.value(), in.dim());
const size_t size = in.size(d);

Expand Down
2 changes: 1 addition & 1 deletion backends/cadence/hifi/operators/operators.h
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ ::executorch::aten::Tensor& div_out_mode(
::executorch::runtime::KernelRuntimeContext& ctx,
const ::executorch::aten::Tensor& a,
const ::executorch::aten::Tensor& b,
::executorch::aten::optional<::executorch::aten::string_view> mode,
std::optional<std::string_view> mode,
::executorch::aten::Tensor& out);

} // namespace native
Expand Down
4 changes: 2 additions & 2 deletions backends/cadence/hifi/operators/tests/test_op_div.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -25,16 +25,16 @@ namespace HiFi {
namespace native {
namespace {

using ::executorch::aten::optional;
using ::executorch::aten::Scalar;
using ::executorch::aten::ScalarType;
using ::executorch::aten::string_view;
using ::executorch::aten::Tensor;
using ::executorch::aten::TensorImpl;
using ::executorch::runtime::Error;
using ::executorch::runtime::KernelRuntimeContext;
using ::executorch::runtime::runtime_init;
using ::executorch::runtime::testing::TensorFactory;
using std::optional;
using std::string_view;

class HiFiDivTest : public OperatorTest {
public:
Expand Down
2 changes: 1 addition & 1 deletion backends/cadence/reference/operators/operators.h
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ using ::executorch::runtime::getLeadingDims;
inline __attribute__((always_inline)) void linear_(
const ::executorch::aten::Tensor& input,
const ::executorch::aten::Tensor& weight,
const ::executorch::aten::optional<::executorch::aten::Tensor>& bias,
const std::optional<::executorch::aten::Tensor>& bias,
::executorch::aten::Tensor& output) {
const float* __restrict__ input_data = input.const_data_ptr<float>();
const float* __restrict__ weight_data = weight.const_data_ptr<float>();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -13,10 +13,10 @@ namespace impl {
namespace reference {
namespace native {

using ::executorch::aten::optional;
using ::executorch::aten::ScalarType;
using ::executorch::aten::Tensor;
using ::executorch::runtime::KernelRuntimeContext;
using std::optional;

void quantized_fully_connected_out(
__ET_UNUSED KernelRuntimeContext& ctx,
Expand Down
Loading
Loading