Back to Graph
R

RLHF

Concept

Reinforcement Learning from Human Feedback - a technique to align AI models with human preferences.

AITrainingAlignment

Relationship Graph

1 nodes0 edges

Statistics

Importance Score90
Connections0

Relationships

No relationships found