Learn Before
  • Gated Linear Unit (GLU)

GeGLU (GELU-based Gated Linear Unit)

GeGLU is a variant of the Gated Linear Unit (GLU) family of activation functions. It is created by specifying the internal non-linear activation function, σ()\sigma(\cdot), to be the Gaussian Error Linear Unit (GELU). This choice distinguishes GeGLU from other GLU variants.

Image 0

0

1

12 days ago

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related
  • Gated Linear Unit (GLU) Formula

  • GeGLU (GELU-based Gated Linear Unit)

  • SwiGLU (Swish-based Gated Linear Unit)

Learn After
  • GeGLU (GELU-based Gated Linear Unit) Formula

  • Applications of GeGLU in Large Language Models