Multiple Choice

A development team is working on an AI assistant. After its initial training, they find that while the assistant's answers are factually accurate, they are often perceived as blunt or unhelpful. To address this, the team decides to use a process where human evaluators are shown a user's prompt followed by two or more different responses generated by the assistant. Which of the following tasks, given to the human evaluators, would be most effective for refining the model's helpfulness and tone?

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.4 Alignment - Foundations of Large Language Models

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science