Skip to content

Commit f3e8213

Browse files
committed
Fix merge conflicts
1 parent b40191b commit f3e8213

File tree

12 files changed

+4
-113
lines changed

12 files changed

+4
-113
lines changed
Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1 @@
1-
<<<<<<< HEAD
21
6193b30becb1ac7be704cf87b8cb9bf13e7f9689
3-
=======
4-
bbb06c0334a6772b92d24bde54956e675c8c6604
5-
>>>>>>> upstream/main

.ci/docker/common/install_rocm.sh

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -114,12 +114,9 @@ EOF
114114
rm -rf HIP clr
115115
fi
116116

117-
<<<<<<< HEAD
118117
# temporary hipblasLT dependency install
119118
apt install libmsgpackc2
120-
=======
121119
pip_install "git+https://github.com/rocm/composable_kernel@$ROCM_COMPOSABLE_KERNEL_VERSION"
122-
>>>>>>> upstream/main
123120

124121
# Cleanup
125122
apt-get autoclean && apt-get clean
@@ -131,8 +128,8 @@ install_centos() {
131128
yum update -y
132129
yum install -y kmod
133130
yum install -y wget
134-
135-
if [[ $OS_VERSION == 9 ]]; then
131+
132+
if [[ $OS_VERSION == 9 ]]; then
136133
dnf install -y openblas-serial
137134
dnf install -y dkms kernel-headers kernel-devel
138135
else

.ci/docker/requirements-ci.txt

Lines changed: 0 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -112,13 +112,8 @@ ninja==1.11.1.3
112112
#Pinned versions: 1.11.1.3
113113
#test that import: run_test.py, test_cpp_extensions_aot.py,test_determination.py
114114

115-
<<<<<<< HEAD
116115
numba==0.60.0 ; python_version == "3.9"
117116
numba==0.61.2 ; python_version > "3.9"
118-
=======
119-
numba==0.55.2 ; python_version == "3.10" and platform_machine != "s390x"
120-
numba==0.60.0 ; python_version == "3.12" and platform_machine != "s390x"
121-
>>>>>>> upstream/main
122117
#Description: Just-In-Time Compiler for Numerical Functions
123118
#Pinned versions: 0.54.1, 0.49.0, <=0.49.1
124119
#test that import: test_numba_integration.py
@@ -137,14 +132,8 @@ numba==0.60.0 ; python_version == "3.12" and platform_machine != "s390x"
137132
#test_nn.py, test_namedtensor.py, test_linalg.py, test_jit_cuda_fuser.py,
138133
#test_jit.py, test_indexing.py, test_datapipe.py, test_dataloader.py,
139134
#test_binary_ufuncs.py
140-
<<<<<<< HEAD
141135
numpy==2.0.2 ; python_version == "3.9"
142136
numpy==2.1.2 ; python_version > "3.9"
143-
=======
144-
numpy==1.22.4; python_version == "3.10"
145-
numpy==1.26.2; python_version == "3.11" or python_version == "3.12"
146-
numpy==2.1.2; python_version >= "3.13"
147-
>>>>>>> upstream/main
148137

149138
pandas==2.2.3
150139

CMakeLists.txt

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -896,8 +896,6 @@ cmake_dependent_option(
896896
"USE_CUDA OR USE_ROCM"
897897
OFF)
898898

899-
<<<<<<< HEAD
900-
=======
901899
IF(USE_FBGEMM_GENAI AND USE_ROCM AND NOT "gfx942" IN_LIST PYTORCH_ROCM_ARCH)
902900
message(WARNING "Unsupported ROCM arch for FBGEMM GenAI, will set USE_FBGEMM_GENAI to OFF")
903901
set(USE_FBGEMM_GENAI off)
@@ -909,7 +907,6 @@ if(USE_CUDA AND "$ENV{TORCH_CUDA_ARCH_LIST}" MATCHES "10.0" AND CMAKE_CUDA_COMPI
909907
set(USE_FBGEMM_GENAI ON)
910908
endif()
911909

912-
>>>>>>> upstream/main
913910
# CAVEAT: Again, Flash Attention2 will error while building for sm52 while Mem
914911
# Eff Attention won't
915912
cmake_dependent_option(

aten/src/ATen/native/Normalization.cpp

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -671,13 +671,9 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, int64_t> _batch_norm_impl_index(
671671
std::cout << "PYTORCH_MIOPEN_EXTRA_LOGGING: ********************* _batch_norm_impl_index (calling miopen_batch_norm)" << std::endl;
672672
return std::tuple_cat(
673673
at::miopen_batch_norm(
674-
<<<<<<< HEAD
675-
input.contiguous(input.suggest_memory_format()), weight.contiguous(), bias.contiguous(),
676-
=======
677674
input.contiguous(input.suggest_memory_format()),
678675
weight.contiguous(),
679676
bias.contiguous(),
680-
>>>>>>> upstream/main
681677
running_mean.defined() ? running_mean.contiguous() : running_mean,
682678
running_var.defined() ? running_var.contiguous() : running_var,
683679
training, momentum, eps),

aten/src/ATen/native/miopen/BatchNorm_miopen.cpp

Lines changed: 0 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -103,11 +103,7 @@ std::tuple<Tensor, Tensor, Tensor> miopen_batch_norm(
103103
mode = miopenBNSpatial;
104104
}
105105

106-
<<<<<<< HEAD
107-
auto output_t = at::empty(input->sizes(), input->options(), input->suggest_memory_format());
108-
=======
109106
auto output_t = at::empty_like(input_t, input_t.options(), input_t.suggest_memory_format());
110-
>>>>>>> upstream/main
111107
TensorArg output{ output_t, "output", 0 };
112108

113109
auto handle = getMiopenHandle();
@@ -180,18 +176,10 @@ std::tuple<Tensor, Tensor, Tensor> miopen_batch_norm_backward(
180176

181177
auto grad_output_contig =
182178
grad_output_t.contiguous(input_t.suggest_memory_format());
183-
<<<<<<< HEAD
184-
TensorArg input{ input_t, "input", 1 },
185-
grad_output{ grad_output_contig, "grad_output", 2 },
186-
weight{ weight_t, "weight", 3 },
187-
save_mean{ save_mean_t, "save_mean", 4 },
188-
save_var{ save_var_t, "save_var", 5 };
189-
=======
190179
TensorArg input{input_t, "input", 1},
191180
grad_output{grad_output_contig, "grad_output", 2},
192181
weight{weight_t, "weight", 3}, save_mean{save_mean_t, "save_mean", 4},
193182
save_var{save_var_t, "save_var", 5};
194-
>>>>>>> upstream/main
195183
CheckedFrom c = "miopen_batch_norm_backward";
196184

197185
checkAllDefined(c, {input, grad_output, weight, save_mean, save_var});
@@ -203,13 +191,9 @@ std::tuple<Tensor, Tensor, Tensor> miopen_batch_norm_backward(
203191
}
204192
checkAllSameType(c, {input, grad_output});
205193
checkAllSameType(c, {weight, save_mean, save_var});
206-
<<<<<<< HEAD
207-
checkAllContiguous(c, {save_mean, save_var});
208-
=======
209194
// TODO: is weight required to be contiguous?
210195
checkAllContiguous(c, {save_mean, save_var});
211196
// TODO: TensorArg check should start handle memory format
212-
>>>>>>> upstream/main
213197
TORCH_CHECK(input->is_contiguous(input->suggest_memory_format()));
214198
TORCH_CHECK(grad_output->is_contiguous(input->suggest_memory_format()));
215199
checkDimRange(c, input, 2, 6 /* exclusive */);
@@ -226,12 +210,7 @@ std::tuple<Tensor, Tensor, Tensor> miopen_batch_norm_backward(
226210
mode = miopenBNSpatial;
227211
}
228212

229-
<<<<<<< HEAD
230-
auto grad_input_t = at::empty(
231-
input->sizes(), input->options(), input->suggest_memory_format());
232-
=======
233213
auto grad_input_t = at::empty(input->sizes(), input->options(), input->suggest_memory_format());
234-
>>>>>>> upstream/main
235214
auto grad_weight_t = at::empty(weight->sizes(), weight->options());
236215
auto grad_bias_t = at::empty(weight->sizes(), weight->options());
237216

requirements-build.txt

Lines changed: 0 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,5 @@
11
# Build System requirements
22
setuptools>=70.1.0,<80.0 # setuptools develop deprecated on 80.0
3-
<<<<<<< HEAD
43
cmake>=3.31.4
54
ninja==1.11.1.3
65
numpy==2.0.2 ; python_version == "3.9"
@@ -10,14 +9,4 @@ pyyaml==6.0.2
109
requests==2.32.4
1110
six==1.17.0 # dependency chain: NNPACK -> PeachPy -> six
1211
typing-extensions==4.14.1
13-
=======
14-
cmake>=3.27
15-
ninja
16-
numpy
17-
packaging
18-
pyyaml
19-
requests
20-
six # dependency chain: NNPACK -> PeachPy -> six
21-
typing-extensions>=4.10.0
2212
pip # not technically needed, but this makes setup.py invocation work
23-
>>>>>>> upstream/main

test/nn/test_convolution.py

Lines changed: 0 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -50,12 +50,6 @@
5050
parametrize as parametrize_test,
5151
run_tests,
5252
set_default_dtype,
53-
<<<<<<< HEAD
54-
skipIfRocm,
55-
skipIfNotMiopenSuggestNHWC,
56-
skipIfRocmVersionLessThan,
57-
=======
58-
>>>>>>> upstream/main
5953
subtest,
6054
TEST_SCIPY,
6155
TEST_WITH_ROCM,
@@ -4033,16 +4027,9 @@ def test_conv_double_backward_strided_with_3D_input_and_weight(self, device):
40334027

40344028
@skipCUDAIfRocm
40354029
@onlyCUDA
4036-
<<<<<<< HEAD
4037-
@largeTensorTest('40GB')
4038-
@largeTensorTest('24GB', 'cpu')
4039-
# Skipped for ROCm temp - https://ontrack-internal.amd.com/browse/SWDEV-383635
4040-
@skipIfRocm
4041-
=======
40424030
@largeTensorTest("40GB")
40434031
@largeTensorTest("24GB", "cpu")
40444032
@tf32_on_and_off(0.005)
4045-
>>>>>>> upstream/main
40464033
def test_conv3d_64bit_indexing(self, device):
40474034
x = torch.rand(1, 32, 512, 512, 256)
40484035
m = torch.nn.Conv3d(32, 1, kernel_size=1, padding=0, stride=1, bias=False)

test/test_binary_ufuncs.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1481,11 +1481,7 @@ def to_np(value):
14811481
elif torch.can_cast(torch.result_type(base, exponent), base.dtype):
14821482
actual2 = actual.pow_(exponent)
14831483
self.assertEqual(actual, expected.to(actual))
1484-
<<<<<<< HEAD
1485-
self.assertEqual(actual2, expected.to(actual))
1486-
=======
14871484
self.assertEqual(actual2, expected.to(actual2))
1488-
>>>>>>> upstream/main
14891485
else:
14901486
self.assertRaisesRegex(
14911487
RuntimeError,

test/test_nn.py

Lines changed: 0 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -5199,24 +5199,6 @@ def test_batchnorm_nhwc_cuda(self):
51995199
name_fn=lambda f, b, m, t: f"{f}_vs_{b}{'_mixed' if m else ''}_{dtype_name(t)}"
52005200
)
52015201
def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype):
5202-
<<<<<<< HEAD
5203-
if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16":
5204-
self.skipTest("3D float16 NCHW train failed on CUDA and ROCm due to Native batchnorm accuracy issue SWDEV-541024")
5205-
if torch.version.hip:
5206-
if self._testMethodName in ("test_batchnorm_2D_train_NHWC_vs_NCHW_mixed_bfloat16",
5207-
"test_batchnorm_2D_train_NCHW_vs_cpu_mixed_bfloat16",
5208-
"test_batchnorm_3D_train_NHWC_vs_NCHW_mixed_bfloat16",
5209-
"test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16"
5210-
) and _get_torch_rocm_version() < (6, 4):
5211-
# NCHW bfloat16 path uses native kernels for rocm<=6.3
5212-
# train failed on rocm<=6.3 due to native tolerance issue SWDEV-507600
5213-
self.skipTest("bfloat16 NHWC train failed on ROCm <= 6.3")
5214-
5215-
if self._testMethodName in ("test_batchnorm_2D_train_NCHW_vs_native_mixed_bfloat16",
5216-
"test_batchnorm_3D_train_NCHW_vs_native_mixed_bfloat16"
5217-
) and _get_torch_rocm_version() >= (6, 4):
5218-
self.skipTest("bfloat16 NCHW train failed due to native tolerance issue SWDEV-507600")
5219-
=======
52205202
if torch.version.cuda:
52215203
if self._testMethodName in ("test_batchnorm_2D_train_NCHW_vs_cpu_mixed_bfloat16",
52225204
"test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16",
@@ -5244,7 +5226,6 @@ def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype):
52445226

52455227
if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16":
52465228
self.skipTest("3D float16 NCHW train failed on ROCm")
5247-
>>>>>>> upstream/main
52485229

52495230
if dims == 3 and memory_format in ("NHWC", "NCHW"):
52505231
memory_format = memory_format + "3D"

0 commit comments

Comments
 (0)