RankFloRankFlo

ai

RLHF

Reinforcement learning from human feedback.

Definition

RLHF (Reinforcement Learning from Human Feedback) is a training technique that uses human ratings to align model outputs with human preferences. Made ChatGPT possible.

Put RLHF to work

Build, publish, and rank your content with RankFlo. Free to start.

Start for free