Notes from course: Generative AI with Large Language Models - Week 3

Notes from Week 1
Notes from Week 2
Reinforcement Learning from Human Feedback (RLHF)
RLHF is essentially a fine-tuning with human feedback which helps to better align models with human prefer...