Inference Process with a Fine-Tuned Model
Once a model's parameters have been optimized through fine-tuning, the resulting model, denoted as , can be used for inference on new, unseen data. For instance, in a text classification task, the new text is first tokenized into a sequence of tokens, represented as . This token sequence is then fed into the fine-tuned model, which processes it to generate a probability distribution over the possible classes.

0
1
References
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.2 Generative Models - Foundations of Large Language Models
Related
Inference Process with a Fine-Tuned Model
Fine-Tuning Objective Function
Complexity and Factors of BERT Fine-Tuning
Formula for Integrating a Prediction Network with a Pre-trained BERT Model
A team of developers starts with a large, general-purpose language model that was trained on a vast corpus of internet text. Their goal is to create a specialized tool that can classify legal documents into specific categories (e.g., 'contract', 'litigation', 'intellectual property'). To do this, they add a new classification component to the model and then train the entire system on a curated, labeled dataset of legal documents. Which statement best analyzes the state of the model's parameters after this training process is successfully completed?
Diagnosing a Fine-Tuning Failure
A machine learning engineer wants to adapt a large, general-purpose language model to perform sentiment analysis on customer reviews. Arrange the following steps in the correct chronological order to successfully specialize the model for this new task.
Different standards for tokenization
Inference Process with a Fine-Tuned Model
Example of Tokenization into Words and Punctuation
Example of Word and Punctuation Tokenization
Methods of Tokenization
A language model is given the sentence: 'The researcher is studying neuroplasticity.' It processes the sentence using two different methods, resulting in two different sequences of tokens.
Method A:
['The', 'researcher', 'is', 'studying', 'neuroplasticity', '.']Method B:['The', 'researcher', 'is', 'study', 'ing', 'neuro', 'plasticity', '.']Assuming the model has never encountered the word 'neuroplasticity' during its training but has seen words like 'neuroscience' and 'plasticity' separately, which method is more advantageous for helping the model understand the new word, and why?
Tokenization Strategies
Evaluating Tokenization for a Specialized Chatbot
Inference Process with a Fine-Tuned Model
A probabilistic model is designed to classify news headlines into one of four categories: 'Business', 'Technology', 'Sports', or 'Health'. For a new headline, the model outputs the following probability distribution:
- Business: 0.12
- Technology: 0.25
- Sports: 0.55
- Health: 0.08
Based on the most common decision rule for converting these probabilities into a single prediction, which category will be assigned to the headline?
Customer Support Ticket Classification
Explaining the Prediction Rule
Inference Process with a Fine-Tuned Model
Probability Distribution Formula for an Encoder-Softmax Language Model
A language model has been trained on a large corpus of English text. When given the sentence 'The chef carefully seasoned the soup with a pinch of ____.', which of the following best represents the direct output the model calculates for the blank position?
Evaluating Sentence Probability
Impact of Training Data on Probability
Learn After
Example of Text for Polarity Classification
Output Formula for a Polarity Classification Model
Example of a Prediction in Polarity Classification
A model, which has been specialized for a text classification task, processes a new input and produces the following probability distribution over three possible classes:
{"Bug Report": 0.15, "Feature Request": 0.75, "General Inquiry": 0.10}. Based on this output, what is the model's final prediction?A language model has been specialized to classify customer support tickets into categories like 'Billing Issue', 'Technical Support', or 'Account Question'. Arrange the following steps in the correct sequence to describe how this model would process a new, unseen customer ticket to make a prediction.
Applying a Specialized Language Model