Your browser doesn't support HTML5 audio
| Definition | : | Reinforcement Learning from Human Feedback |
| Category | : | Computing » Artificial Intelligence |
| Country/ Region |
: | Worldwide
|
| Popularity | : |
|
| Type | : |
Initialism
|
Reinforcement Learning from Human Feedback (RLHF) is a type of machine learning method where an agent learns to make decisions by receiving feedback from a human in the form of rewards or punishments.
Note:
An agent is a software or algorithm that is designed to take actions in an environment in order to achieve a specific goal or set of goals.
The full form of RLHF is Reinforcement Learning from Human Feedback
Reinforcement Learning from Human Feedback