Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -112,6 +112,12 @@ This model is intended for researchers, developers, and organizations seeking a
112
  ## Training Data
113
  The `Locutusque/Hyperion-3.0-Mistral-7B-DPO` model was fine-tuned on a carefully curated dataset of 20,000 preference pairs, where 4,000 examples were used to fine-tune. These examples were generated by GPT-4 to ensure the highest quality and relevance across various domains, including programming, medical texts, mathematical problems, and reasoning tasks. The training data was further optimized using Direct Preference Optimization (DPO) to align the model's outputs with human preferences and improve overall performance.
114
 
 
 
 
 
 
 
115
  ## Evaluation Results
116
  mmlu flan cot 5-shot
117
 
 
112
  ## Training Data
113
  The `Locutusque/Hyperion-3.0-Mistral-7B-DPO` model was fine-tuned on a carefully curated dataset of 20,000 preference pairs, where 4,000 examples were used to fine-tune. These examples were generated by GPT-4 to ensure the highest quality and relevance across various domains, including programming, medical texts, mathematical problems, and reasoning tasks. The training data was further optimized using Direct Preference Optimization (DPO) to align the model's outputs with human preferences and improve overall performance.
114
 
115
+ ## Quants
116
+
117
+ ExLlamaV2: https://huggingface.co/bartowski/Hyperion-3.0-Mistral-7B-DPO-exl2
118
+
119
+ GGUF: https://huggingface.co/bartowski/Hyperion-3.0-Mistral-7B-DPO-GGUF
120
+
121
  ## Evaluation Results
122
  mmlu flan cot 5-shot
123