News

  • 13 Feb 2025: I have updated my website for the first time in last two years :D

2025

RLHF Basics

less than 1 minute read

The goal of Reinforcement Learning from Human Feedback (RLHF) is aligning LLMs with human expectation. This post explains some of its basics.

Back to Top ↑