@@ -21,8 +21,7 @@ namespace math {
21
21
template <typename T>
22
22
__global__ void KernelUnpool2dMax (const int nthreads, const T* input_data,
23
23
const int * indices_data,
24
- const int input_height,
25
- const int input_width,
24
+ const int input_height, const int input_width,
26
25
const int channels, T* output_data,
27
26
const int output_height,
28
27
const int output_width) {
@@ -71,8 +70,8 @@ template <typename T>
71
70
class Unpool2dMaxFunctor <platform::GPUPlace, T> {
72
71
public:
73
72
void operator ()(const platform::DeviceContext& context,
74
- const framework::Tensor& input,
75
- const framework::Tensor& indices, framework::Tensor* output) {
73
+ const framework::Tensor& input,
74
+ const framework::Tensor& indices, framework::Tensor* output) {
76
75
const int batch_size = input.dims ()[0 ];
77
76
const int input_height = input.dims ()[2 ];
78
77
const int input_width = input.dims ()[3 ];
@@ -86,10 +85,10 @@ class Unpool2dMaxFunctor<platform::GPUPlace, T> {
86
85
int grid = (input.numel () + threads - 1 ) / threads;
87
86
KernelUnpool2dMax<
88
87
T><<<grid, threads, 0 ,
89
- reinterpret_cast <const platform::CUDADeviceContext&>(context)
90
- .stream()>>> (input.numel (), input_data, indices_data,
91
- input_height, input_width, output_channels,
92
- output_data, output_height, output_width);
88
+ reinterpret_cast <const platform::CUDADeviceContext&>(context)
89
+ .stream()>>> (input.numel (), input_data, indices_data,
90
+ input_height, input_width, output_channels,
91
+ output_data, output_height, output_width);
93
92
}
94
93
};
95
94
/*
@@ -119,11 +118,11 @@ class Unpool2dMaxGradFunctor<platform::GPUPlace, T> {
119
118
int grid = (input.numel () + threads - 1 ) / threads;
120
119
KernelUnpool2dMaxGrad<
121
120
T><<<grid, threads, 0 ,
122
- reinterpret_cast <const platform::CUDADeviceContext&>(context)
123
- .stream()>>> (input.numel (), input_data, indices_data,
124
- input_height, input_width, output_channels,
125
- output_data, output_grad_data, output_height,
126
- output_width, input_grad_data);
121
+ reinterpret_cast <const platform::CUDADeviceContext&>(context)
122
+ .stream()>>> (input.numel (), input_data, indices_data,
123
+ input_height, input_width, output_channels,
124
+ output_data, output_grad_data, output_height,
125
+ output_width, input_grad_data);
127
126
}
128
127
};
129
128
template class Unpool2dMaxGradFunctor <platform::GPUPlace, float >;
0 commit comments