You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
### Understanding the PReLU (Parametric ReLU) Activation Function
2
+
3
+
The PReLU (Parametric Rectified Linear Unit) is an advanced variant of the ReLU activation function that introduces a learnable parameter for negative inputs. This makes it more flexible than standard ReLU and helps prevent the "dying ReLU" problem.
4
+
5
+
#### Mathematical Definition
6
+
7
+
The PReLU function is defined as:
8
+
9
+
$$
10
+
PReLU(x) = \begin{cases}
11
+
x & \text{if } x > 0 \\
12
+
\alpha x & \text{otherwise}
13
+
\end{cases}
14
+
$$
15
+
16
+
Where:
17
+
- $x$ is the input value
18
+
- $\alpha$ is a learnable parameter (typically initialized to a small value like 0.25)
19
+
20
+
#### Key Characteristics
21
+
22
+
1.**Adaptive Slope**: Unlike ReLU which has a zero slope for negative inputs, PReLU learns the optimal negative slope parameter ($\alpha$) during training.
23
+
24
+
2.**Output Range**:
25
+
- For $x > 0$: Output equals input ($y = x$)
26
+
- For $x \leq 0$: Output is scaled by $\alpha$ ($y = \alpha x$)
27
+
28
+
3.**Advantages**:
29
+
- Helps prevent the "dying ReLU" problem
30
+
- More flexible than standard ReLU
31
+
- Can improve model performance through learned parameter
32
+
- Maintains the computational efficiency of ReLU
33
+
34
+
4.**Special Cases**:
35
+
- When $\alpha = 0$, PReLU becomes ReLU
36
+
- When $\alpha = 1$, PReLU becomes a linear function
37
+
- When $\alpha$ is small (e.g., 0.01), PReLU behaves similarly to Leaky ReLU
38
+
39
+
PReLU is particularly useful in deep neural networks where the optimal negative slope might vary across different layers or channels.
0 commit comments