1212# See the License for the specific language governing permissions and
1313# limitations under the License.
1414# ==============================================================================
15- """Implements GeLU activation."""
15+ """Implements GELU activation."""
1616
1717from __future__ import absolute_import
1818from __future__ import division
2323
2424
2525@tf .keras .utils .register_keras_serializable (package = 'Addons' )
26- class GeLU (tf .keras .layers .Layer ):
26+ class GELU (tf .keras .layers .Layer ):
2727 """Gaussian Error Linear Unit.
2828
2929 A smoother version of ReLU generally used
@@ -40,7 +40,7 @@ class GeLU(tf.keras.layers.Layer):
4040 """
4141
4242 def __init__ (self , approximate = True , ** kwargs ):
43- super (GeLU , self ).__init__ (** kwargs )
43+ super (GELU , self ).__init__ (** kwargs )
4444 self .approximate = approximate
4545 self .supports_masking = True
4646
@@ -49,7 +49,7 @@ def call(self, inputs):
4949
5050 def get_config (self ):
5151 config = {'approximate' : self .approximate }
52- base_config = super (GeLU , self ).get_config ()
52+ base_config = super (GELU , self ).get_config ()
5353 return dict (list (base_config .items ()) + list (config .items ()))
5454
5555 def compute_output_shape (self , input_shape ):
0 commit comments