diff --git a/keras_nlp/models/t5/t5_transformer_layer.py b/keras_nlp/models/t5/t5_transformer_layer.py index f70010e42d..1750232f98 100644 --- a/keras_nlp/models/t5/t5_transformer_layer.py +++ b/keras_nlp/models/t5/t5_transformer_layer.py @@ -123,7 +123,7 @@ def call( x = self.layer_norm(x) if self.use_gated_activation: hidden_activation = self.input_projector(x) - hidden_linear = self.gate_projector(hidden_states) + hidden_linear = self.gate_projector(x) x = hidden_activation * hidden_linear else: x = self.input_projector(x)