Learn Before
Numerical Overflow in Softmax Function
When calculating the softmax function, , computational risks arise due to exponential operations. If some input logits, , are very large positive numbers, computing can produce values that exceed the maximum limit of certain data types (such as for single-precision floating-point numbers). This phenomenon is known as numerical overflow, and it leads to mathematical instability because the resulting predicted probabilities become undefined.
0
1
Tags
D2L
Dive into Deep Learning @ D2L
Related
Pros and Cons of Softmax Function
Softmax Regression (Activation)
Parameterized Softmax Layer
Plackett-Luce Selection Probability Formula
Conditional Probability Formula for Autoregressive Models using Softmax
A neural network's final layer produces the raw output scores (logits)
[2.0, 1.0, 0.1]for three possible classes. To convert these scores into class probabilities, a function is applied that first exponentiates each score and then normalizes these new values by dividing each by their sum. What is the resulting probability distribution? (Values are rounded to three decimal places).A function is used to convert a vector of raw, unnormalized scores
z = [z_1, z_2, ..., z_K]into a probability distribution. This function operates by first applying the standard exponential function to each score and then normalizing these new values by dividing each by their sum. If a constant valueCis added to every score in the input vectorz, resulting in a new vectorz' = [z_1+C, z_2+C, ..., z_K+C], how will the resulting output probability distribution be affected?Consider two input vectors of raw scores (logits) for a 3-class classification problem: Vector A =
[1, 2, 3]and Vector B =[1, 5, 10]. Both vectors are passed through a function that exponentiates each score and then normalizes the results by dividing by their sum. How will the resulting probability distribution for Vector B compare to the one for Vector A?Youâre reviewing an internal evaluation script tha...
Your team is building an internal tool that ranks ...
Youâre reviewing an internal LLM evaluation pipeli...
Reconciling Training Log-Likelihood with Inference-Time Sequence Selection
Explaining a Counterintuitive Decoding Outcome Using Softmax, Next-Token Conditionals, and Sequence Log-Probability
Diagnosing a âHigh-Confidence Wrong Tokenâ Bug in Autoregressive Scoring
Investigating a Production Scoring Bug: Softmax Normalization vs. Autoregressive Sequence Log-Probability
Design a Correct Sequence-Scoring Function for Autoregressive LLM Outputs
Root-Cause Analysis: Why a âMore Likelyâ Token-by-Token Completion Loses on Total Sequence Score
Auditing a Candidate Completion Using Softmax Next-Token Probabilities and Autoregressive Log-Probability
Derivative of Softmax Cross-Entropy Loss with Respect to Logits
Numerical Overflow in Softmax Function