Skip to content

Commit 2df04ca

Browse files
using old patch for vec error
1 parent ee92a60 commit 2df04ca

File tree

6 files changed

+8
-16
lines changed

6 files changed

+8
-16
lines changed

.github/workflows/_binary-build-linux.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -273,7 +273,7 @@ jobs:
273273
-w / \
274274
"${DOCKER_IMAGE}"
275275
)
276-
#docker exec -t -w "${PYTORCH_ROOT}" "${container_name}" bash -c "bash .circleci/scripts/binary_populate_env.sh"
276+
docker exec -t -w "${PYTORCH_ROOT}" "${container_name}" bash -c "bash .circleci/scripts/binary_populate_env.sh"
277277
if [[ ${BUILD_ENVIRONMENT} == *"aarch64"* ]]; then
278278
docker exec -t "${container_name}" bash -c "source ${BINARY_ENV_FILE} && bash /pytorch/.ci/aarch64_linux/aarch64_ci_build.sh"
279279
else

aten/src/ATen/cpu/vec/vec256/vsx/vec256_common_vsx.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,9 +14,9 @@
1414
#include <ATen/cpu/vec/vec256/vsx/vec256_qint8_vsx.h>
1515
#include <ATen/cpu/vec/vec256/vsx/vec256_quint8_vsx.h>
1616

17-
#include <ATen/cpu/vec/vec256/vsx/vec256_complex_double_vsx.h>
18-
#include <ATen/cpu/vec/vec256/vsx/vec256_complex_float_vsx.h>
1917

18+
#include <ATen/cpu/vec/vec256/vsx/vec256_complex_float_vsx.h>
19+
#include <ATen/cpu/vec/vec256/vsx/vec256_complex_double_vsx.h>
2020

2121
#include <ATen/cpu/vec/vec256/vsx/vec256_bfloat16_vsx.h>
2222

aten/src/ATen/cpu/vec/vec256/vsx/vec256_complex_double_vsx.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -478,7 +478,7 @@ class Vectorized<ComplexDbl> {
478478
this->store(tmp1);
479479
b.store(tmp2);
480480

481-
for (const auto i : c10::irange(Vectorized<c10::complex<double>>::size())) {
481+
for (const auto i : c10::irange(Vectorized<c10::complex<float>>::size())) {
482482
out[i] = tmp1[i] / tmp2[i];
483483
}
484484
return loadu(out);

aten/src/ATen/native/Blas.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -296,7 +296,7 @@ _scaled_mm_out_cpu(const Tensor& mat1, const Tensor& mat2,
296296
std::optional<c10::ScalarType> out_dtype,
297297
bool use_fast_accum,
298298
Tensor& out) {
299-
#if AT_MKLDNN_ENABLED() && !defined(__powerpc__) && !defined(__powerpc64__) && !defined(__PPC__) && !defined(__ppc__)
299+
#if AT_MKLDNN_ENABLED() && !defined(__powerpc__)
300300
if (at::globalContext().userEnabledMkldnn()) {
301301
bool mixed_dtype = mat1.scalar_type() != mat2.scalar_type();
302302
if ((!mixed_dtype && cpuinfo_has_x86_amx_int8()) ||

aten/src/ATen/native/mkldnn/Conv.cpp

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -161,13 +161,8 @@ static bool mkldnn_conv_enabled_fpmath_mode_bf16(){
161161
}
162162

163163
static bool mkldnn_conv_enabled_fpmath_mode_tf32(){
164-
#if defined(__powerpc64__)
165-
// TF32 is not supported or relevant on Power — skip
166-
return false;
167-
#else
168-
return at::globalContext().float32Precision("mkldnn", "conv") == "tf32" &&
169-
cpuinfo_has_x86_amx_fp16(); // or whatever the original intent was
170-
#endif
164+
return at::globalContext().float32Precision("mkldnn", "conv") == "tf32" &&
165+
cpuinfo_has_x86_amx_fp16();
171166
}
172167

173168
static inline at::MemoryFormat mkldnn_convolution_memory_format(int64_t dims, bool is_channels_last) {

cmake/Modules/FindMKLDNN.cmake

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -85,12 +85,9 @@ IF(NOT MKLDNN_FOUND)
8585
ENDIF(NOT APPLE AND NOT WIN32 AND NOT BUILD_LITE_INTERPRETER)
8686

8787
IF(EXISTS "${MKLDNN_ROOT}/include/oneapi/dnnl/dnnl_ukernel.hpp")
88-
IF(CPU_POWER)
89-
SET(DNNL_EXPERIMENTAL_UKERNEL OFF CACHE BOOL "" FORCE)
90-
ELSE()
9188
MESSAGE("-- Will build oneDNN UKERNEL")
9289
SET(DNNL_EXPERIMENTAL_UKERNEL ON CACHE BOOL "" FORCE)
93-
ENDIF()
90+
9491
ENDIF(EXISTS "${MKLDNN_ROOT}/include/oneapi/dnnl/dnnl_ukernel.hpp")
9592

9693
FIND_PACKAGE(BLAS)

0 commit comments

Comments
 (0)