We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent e8c32e8 commit 26faeb5Copy full SHA for 26faeb5
official/nlp/modeling/layers/transformer.py
@@ -142,10 +142,8 @@ def build(self, input_shape):
142
kernel_constraint=self._kernel_constraint,
143
bias_constraint=self._bias_constraint,
144
name="intermediate")
145
- # Use float32 in intermediate gelu activation for numeric stability.
146
- # TODO(b/149117297): investigate gelu numeric stability.
147
self._intermediate_activation_layer = tf.keras.layers.Activation(
148
- self._intermediate_activation, dtype=tf.float32)
+ self._intermediate_activation)
149
self._output_dense = dense_einsum.DenseEinsum(
150
output_shape=hidden_size,
151
kernel_initializer=self._kernel_initializer,
0 commit comments