Concept

Reward in Reinforcement Learning

In reinforcement learning, a reward is a signal sent from the environment to the agent that provides feedback on an action's success. This feedback, which can be positive or negative, guides the agent's learning process by indicating the desirability of the actions taken. The agent's primary objective is to modify its policy over time to maximize the cumulative reward it receives.

0

3

Updated 2025-10-06

Tags

Data Science

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences