Skip to content
Embedding LabsEmbedding Labs
Embedding Labs

Reinforcement Learning from Human Feedback

Zurück zur Bildung
Level: FortgeschrittenDauer: 19:39Quelle: Google Cloud Tech
RLHFAlignmentTraining

Zusammenfassung

An in-depth exploration of Reinforcement Learning from Human Feedback, the technique that makes language models helpful and harmless. The video covers the three-stage process: supervised fine-tuning, reward model training on human preferences, and policy optimization using PPO. Learn how RLHF enables models to understand nuanced human values, producing outputs that are genuinely useful and aligned with user intent.

Quelle

Quelle:Google Cloud Tech
Dauer:19:39
Level:Fortgeschritten
Topics:
RLHFAlignmentTraining

Transkript