Multiple Choice

An engineer is training two reinforcement learning agents (Agent A and Agent B) on the same task using a policy gradient method. The environment has a wide range of possible total rewards, from highly negative to highly positive. Agent A's learning algorithm directly uses the total reward received after each episode to update its policy. Agent B's algorithm first subtracts a constant value (equal to the average total reward observed so far) from the total reward before using it for the update. What is the most likely difference in the training process between Agent A and Agent B?

0

1

Updated 2025-09-29

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science