Notes on Reinforcement Learning for Human Feedback

Reinforcement Learning for Human Feedback (RLHF) is the concept with powers recent models like ChatGPT. In my notes, I am covering resources I found helpful to get started with RLHF.

Paper

Introduction

Code