Learn Before
Gated Linear Unit (GLU)
GeGLU (GELU-based Gated Linear Unit)
GeGLU is a variant of the Gated Linear Unit (GLU) family of activation functions. It is created by specifying the internal non-linear activation function, , to be the Gaussian Error Linear Unit (GELU). This choice distinguishes GeGLU from other GLU variants.

0
1
12 days ago
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Gated Linear Unit (GLU) Formula
GeGLU (GELU-based Gated Linear Unit)
SwiGLU (Swish-based Gated Linear Unit)
Learn After
GeGLU (GELU-based Gated Linear Unit) Formula
Applications of GeGLU in Large Language Models