Skip to content

Commit 29a9f9b

Browse files
committed
Refine code format and fix threads number.
1 parent 5a4cdbb commit 29a9f9b

File tree

3 files changed

+41
-41
lines changed

3 files changed

+41
-41
lines changed

paddle/operators/math/detail/activation_functions.h

Lines changed: 28 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -32,25 +32,25 @@ namespace detail {
3232
namespace forward {
3333

3434
template <typename T>
35-
DEVICE T linear(const T a) {
35+
DEVICE T Identity(const T a) {
3636
return a;
3737
}
3838

3939
template <typename T>
40-
DEVICE T relu(const T a) {
40+
DEVICE T Relu(const T a) {
4141
return a > static_cast<T>(0.0) ? a : static_cast<T>(0.0);
4242
}
4343

4444
template <typename T>
45-
DEVICE T sigmoid(const T a) {
45+
DEVICE T Sigmoid(const T a) {
4646
const T min = SIGMOID_THRESHOLD_MIN;
4747
const T max = SIGMOID_THRESHOLD_MAX;
4848
T tmp = (a < min) ? min : ((a > max) ? max : a);
4949
return static_cast<T>(1.0) / (static_cast<T>(1.0) + exp(-tmp));
5050
}
5151

5252
template <typename T>
53-
DEVICE T tanh(const T a) {
53+
DEVICE T Tanh(const T a) {
5454
T tmp = -2.0 * a;
5555
tmp = (tmp > EXP_MAX_INPUT) ? EXP_MAX_INPUT : tmp;
5656
return (2.0 / (1.0 + exp(tmp))) - 1.0;
@@ -61,22 +61,22 @@ DEVICE T tanh(const T a) {
6161
namespace backward {
6262

6363
template <typename T>
64-
DEVICE T linear(const T a, const T b) {
64+
DEVICE T Identity(const T a, const T b) {
6565
return a;
6666
}
6767

6868
template <typename T>
69-
DEVICE T relu(const T a, const T b) {
69+
DEVICE T Relu(const T a, const T b) {
7070
return a * (b > 0.0 ? 1.0 : 0.0);
7171
}
7272

7373
template <typename T>
74-
DEVICE T sigmoid(const T a, const T b) {
74+
DEVICE T Sigmoid(const T a, const T b) {
7575
return a * b * (1.0 - b);
7676
}
7777

7878
template <typename T>
79-
DEVICE T tanh(const T a, const T b) {
79+
DEVICE T Tanh(const T a, const T b) {
8080
return a * (1.0 - b * b);
8181
}
8282

@@ -89,20 +89,20 @@ struct Active {
8989
};
9090

9191
static DEVICE Active<float>::Act kActFloat[] = {
92-
&forward::sigmoid<float>, &forward::relu<float>, &forward::tanh<float>,
93-
&forward::linear<float>};
92+
&forward::Sigmoid<float>, &forward::Relu<float>, &forward::Tanh<float>,
93+
&forward::Identity<float>};
9494

9595
static DEVICE Active<float>::ActGrad kActGradFloat[] = {
96-
&backward::sigmoid<float>, &backward::relu<float>, &backward::tanh<float>,
97-
&backward::linear<float>};
96+
&backward::Sigmoid<float>, &backward::Relu<float>, &backward::Tanh<float>,
97+
&backward::Identity<float>};
9898

9999
static DEVICE Active<double>::Act kActDouble[] = {
100-
&forward::sigmoid<double>, &forward::relu<double>, &forward::tanh<double>,
101-
&forward::linear<double>};
100+
&forward::Sigmoid<double>, &forward::Relu<double>, &forward::Tanh<double>,
101+
&forward::Identity<double>};
102102

103103
static DEVICE Active<double>::ActGrad kActGradDouble[] = {
104-
&backward::sigmoid<double>, &backward::relu<double>,
105-
&backward::tanh<double>, &backward::linear<double>};
104+
&backward::Sigmoid<double>, &backward::Relu<double>,
105+
&backward::Tanh<double>, &backward::Identity<double>};
106106

107107
namespace forward {
108108
inline DEVICE float activation(float a, int index) {
@@ -128,29 +128,29 @@ inline DEVICE double activation(double a, double b, int index) {
128128
#ifdef __AVX__
129129
namespace forward {
130130
namespace avx {
131-
__m256 relu(const __m256 a);
132-
__m256 sigmoid(const __m256 a);
133-
__m256 tanh(const __m256 a);
134-
__m256 linear(const __m256 a);
131+
__m256 Relu(const __m256 a);
132+
__m256 Sigmoid(const __m256 a);
133+
__m256 Tanh(const __m256 a);
134+
__m256 Identity(const __m256 a);
135135
} // namespace avx
136136
} // namespace forward
137137

138138
namespace backward {
139139
namespace avx {
140-
__m256 relu(const __m256 a, const __m256 b);
141-
__m256 sigmoid(const __m256 a, const __m256 b);
142-
__m256 tanh(const __m256 a, const __m256 b);
143-
__m256 linear(const __m256 a, const __m256 b);
140+
__m256 Relu(const __m256 a, const __m256 b);
141+
__m256 Sigmoid(const __m256 a, const __m256 b);
142+
__m256 Tanh(const __m256 a, const __m256 b);
143+
__m256 Identity(const __m256 a, const __m256 b);
144144
} // namespace avx
145145
} // namespace backward
146146

147147
static Active<__m256>::Act kActAvx[] = {
148-
&forward::avx::sigmoid, &forward::avx::relu, &forward::avx::tanh,
149-
&forward::avx::linear};
148+
&forward::avx::Sigmoid, &forward::avx::Relu, &forward::avx::Tanh,
149+
&forward::avx::Identity};
150150

151151
static Active<__m256>::ActGrad kActGradAvx[] = {
152-
&backward::avx::sigmoid, &backward::avx::relu, &backward::avx::tanh,
153-
&backward::avx::linear};
152+
&backward::avx::Sigmoid, &backward::avx::Relu, &backward::avx::Tanh,
153+
&backward::avx::Identity};
154154

155155
namespace forward {
156156
inline __m256 activation(__m256 a, int index) { return kActAvx[index](a); }

paddle/operators/math/detail/avx_functions.cc

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -22,61 +22,61 @@ namespace operators {
2222
namespace math {
2323
namespace detail {
2424

25-
__m256 exp(__m256 a) { return exp256_ps(a); }
25+
__m256 Exp(__m256 a) { return exp256_ps(a); }
2626

2727
namespace forward {
2828
namespace avx {
29-
__m256 relu(const __m256 a) {
29+
__m256 Relu(const __m256 a) {
3030
__m256 tmp = _mm256_set1_ps(0.0f);
3131
return _mm256_max_ps(a, tmp);
3232
}
3333

34-
__m256 sigmoid(const __m256 a) {
34+
__m256 Sigmoid(const __m256 a) {
3535
__m256 max = _mm256_set1_ps(SIGMOID_THRESHOLD_MAX);
3636
__m256 min = _mm256_set1_ps(SIGMOID_THRESHOLD_MIN);
3737
__m256 tmp = _mm256_max_ps(a, min);
3838
tmp = _mm256_min_ps(tmp, max);
3939
tmp = _mm256_sub_ps(_mm256_set1_ps(0.0f), tmp);
40-
tmp = exp(tmp);
40+
tmp = Exp(tmp);
4141
tmp = _mm256_add_ps(_mm256_set1_ps(1.0f), tmp);
4242
tmp = _mm256_div_ps(_mm256_set1_ps(1.0f), tmp);
4343
return tmp;
4444
}
4545

46-
__m256 tanh(const __m256 a) {
46+
__m256 Tanh(const __m256 a) {
4747
__m256 max = _mm256_set1_ps(EXP_MAX_INPUT);
4848
__m256 tmp = _mm256_mul_ps(_mm256_set1_ps(-2.0f), a);
4949
tmp = _mm256_min_ps(tmp, max);
50-
tmp = exp(tmp);
50+
tmp = Exp(tmp);
5151
return _mm256_sub_ps(_mm256_div_ps(_mm256_set1_ps(2.0f),
5252
_mm256_add_ps(_mm256_set1_ps(1.0f), tmp)),
5353
_mm256_set1_ps(1.0f));
5454
}
5555

56-
__m256 linear(const __m256 a) { return a; }
56+
__m256 Identity(const __m256 a) { return a; }
5757

5858
} // namespace avx
5959
} // namespace forward
6060

6161
namespace backward {
6262
namespace avx {
63-
__m256 relu(const __m256 a, const __m256 b) {
63+
__m256 Relu(const __m256 a, const __m256 b) {
6464
return _mm256_mul_ps(
6565
a, _mm256_and_ps(_mm256_cmp_ps(b, _mm256_set1_ps(0.0f), _CMP_GT_OS),
6666
_mm256_set1_ps(1.0f)));
6767
}
6868

69-
__m256 sigmoid(const __m256 a, const __m256 b) {
69+
__m256 Sigmoid(const __m256 a, const __m256 b) {
7070
return _mm256_mul_ps(_mm256_mul_ps(a, b),
7171
_mm256_sub_ps(_mm256_set1_ps(1.0f), b));
7272
}
7373

74-
__m256 tanh(const __m256 a, const __m256 b) {
74+
__m256 Tanh(const __m256 a, const __m256 b) {
7575
return _mm256_mul_ps(
7676
a, _mm256_sub_ps(_mm256_set1_ps(1.0f), _mm256_mul_ps(b, b)));
7777
}
7878

79-
__m256 linear(const __m256 a, const __m256 b) { return a; }
79+
__m256 Identity(const __m256 a, const __m256 b) { return a; }
8080
} // namespace avx
8181
} // namespace backward
8282

paddle/operators/math/detail/lstm_gpu_kernel.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -226,9 +226,9 @@ void gpu_lstm_backward(const platform::DeviceContext& context, Op op,
226226
threads = dim3(framePerBlock, 1);
227227
grid = dim3(frameBlocks, 1);
228228
} else {
229-
/* framePerBlock = 32 batchPerBlock = 32 */
229+
/* framePerBlock = 32 batchPerBlock = 16 */
230230
threads = dim3(32, 16);
231-
grid = dim3((frameSize + 32 - 1) / 32, (batchSize + 32 - 1) / 32);
231+
grid = dim3((frameSize + 32 - 1) / 32, (batchSize + 16 - 1) / 16);
232232
}
233233

234234
auto stream =

0 commit comments

Comments
 (0)