You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: NEWS.md
+10-3Lines changed: 10 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,10 +2,17 @@
2
2
3
3
See also [github's page](https://github.com/FluxML/Flux.jl/releases) for a complete list of PRs merged before each release.
4
4
5
-
## v0.15.3
6
-
* Add `WeightNorm` normalization layer.
5
+
## v0.16.0 (15 December 2025)
6
+
This release has a single **breaking change**:
7
7
8
-
## v0.15.0 (December 2024)
8
+
- The recurrent cells `RNNCell`, `LSTMCell`, and `GRUCell` forward has been changed to
9
+
$y_t, state_t = cell(x_t, state_{t-1})$. Previously, it was $state_t = cell(x_t, state_{t-1})$.
10
+
11
+
Other highlights include:
12
+
* Added `WeightNorm` normalization layer.
13
+
* Added `Recurrence` layer, turning a recurrent layer into a layer processing the entire sequence at once.
14
+
15
+
## v0.15.0 (5 December 2024)
9
16
This release includes two **breaking changes**:
10
17
- The recurrent layers have been thoroughly revised. See below and read the [documentation](https://fluxml.ai/Flux.jl/v0.15/guide/models/recurrence/) for details.
11
18
- Flux now defines and exports its own gradient function. Consequently, using gradient in an unqualified manner (e.g., after `using Flux, Zygote`) could result in an ambiguity error.
y = [y; [ht]] # concatenate in non-mutating (AD friendly) way
36
+
yt, ht =rnn_cell(xt, ht)
37
+
y = [y; [yt]] # concatenate in non-mutating (AD friendly) way
38
38
end
39
39
```
40
40
41
41
Notice how the above is essentially a `Dense` layer that acts on two inputs, `xt` and `ht`.
42
-
43
-
The output at each time step, called the hidden state, is used as the input to the next time step and is also the output of the model.
42
+
The result of the forward pass at each time step, is a tuple contening the output `yt` and the updated state `ht`. The updated state is used as an input in next iteration. In the simple case of a vanilla RNN, the
43
+
output and the state are the same. In more complex cells, such as `LSTMCell`, the state can contain multiple arrays.
44
44
45
45
There are various recurrent cells available in Flux, notably `RNNCell`, `LSTMCell` and `GRUCell`, which are documented in the [layer reference](../../reference/models/layers.md). The hand-written example above can be replaced with:
0 commit comments