Reinforcement Learning Human Feedback (RLHF)

What it means - A training method where people tell the AI which answers are better, and the AI learns from that.

Think of it like - A teacher grading essays and telling you how to improve your writing next time.