dpo
Here are 47 public repositories matching this topic...
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
-
Updated
Jun 14, 2024 - Python
Step-aware Preference Optimization: Aligning Preference with Denoising Performance at Each Step
-
Updated
Jun 12, 2024 - Python
CodeUltraFeedback: aligning large language models to coding preferences
-
Updated
May 30, 2024 - Python
This is the DPO Pay plugin for WooCommerce.
-
Updated
May 28, 2024 - PHP
A open-source framework designed to adapt pre-trained Language Models (LLMs), such as Llama, Mistral, and Mixtral, to a wide array of domains and languages.
-
Updated
May 27, 2024 - Python
Examples for using the SiLLM framework for training and running Large Language Models (LLMs) on Apple Silicon
-
Updated
May 17, 2024 - Python
This is the DPO Group plugin for Gravity Forms.
-
Updated
Apr 29, 2024 - PHP
Data and models for the paper "Configurable Safety Tuning of Language Models with Synthetic Preference Data"
-
Updated
Apr 23, 2024 - Python
Various training, inference and validation code and results related to Open LLM's that were pretrained (full or partially) on the Dutch language.
-
Updated
Apr 9, 2024 - Jupyter Notebook
Improve this page
Add a description, image, and links to the dpo topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the dpo topic, visit your repo's landing page and select "manage topics."