Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper
•
2305.18290
•
Published
•
64
Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-α is the first model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). We found that removing the in-built alignment of these datasets boosted performance on MT Bench and made the model more helpful. However, this means that model is likely to generate problematic text when prompted to do so.
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Base model
mistralai/Mistral-7B-v0.1