Let's build better datasets, together!
-
Updated
Jun 6, 2024 - Jupyter Notebook
Let's build better datasets, together!
A curated list of reinforcement learning with human feedback resources (continually updated)
[ICML 2024] Code for the paper "Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases"
Product analytics for AI Assistants
Aligning LLM Agents by Learning Latent Preference from User Edits
The Prism Alignment Project
[CVPR 2024] Code for the paper "Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model"
Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
[ NeurIPS 2023 ] Official Codebase for "Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback"
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
Reinforcement Learning from Human Feedback with 🤗 TRL
A curated list of reinforcement learning with human feedback resources[awesome-RLHF-Turkish] (continually updated)
Implementation of Reinforcement Learning from Human Feedback (RLHF)
Add a description, image, and links to the human-feedback topic page so that developers can more easily learn about it.
To associate your repository with the human-feedback topic, visit your repo's landing page and select "manage topics."