Skip to content

Latest commit

 

History

History

7. RLHF Model

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 

RLHF - Reinforcement Learning using Human Feedback

This directory contains notebooks that can be used to train a reward model, and then fine-tune the LLM using Reinforcement Learning. For a detailed overview of Reward Modeling and RLHF, refer to:

Architecture

RLHF

Notebooks

  1. rewardModelTraining.ipynb : This notebook takes in user preference data as input and trains a model of choice to output scaler reward.

GIF Description

  1. RLHFImplementation.ipynb : This notebook takes a SFT LLM, a reward model and data as input. It then finetunes the LLM using PPO.

GIF Description

References

  1. https://github.com/lvwerra/trl
  2. https://argilla.io/blog/argilla-for-llms/