bullerwins commited on
Commit
3909f47
1 Parent(s): 943b49e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,11 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Reflection-70B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Reflection-70B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Reflection-70B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Reflection-70B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Reflection-70B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Reflection-70B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Reflection-70B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Reflection-70B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
4
+ pipeline_tag: text-generation
5
+ ---
6
+ # Reflection Llama-3.1 70B
7
+
8
+ **Reflection Llama-3.1 70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.**
9
+
10
+ The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
11
+
12
+ You can [try the model here](https://reflection-playground-production.up.railway.app/).
13
+
14
+ ## Benchmarks
15
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60518f3731c5be7f3dd5ebc3/zNs-ZFs0SbnomH7mikiOU.png)
16
+
17
+ All benchmarks tested have been checked for contamination by running [LMSys's LLM Decontaminator](https://github.com/lm-sys/llm-decontaminator). When benchmarking, we isolate the `<output>` and benchmark on solely that section.
18
+
19
+ Trained from Llama 3.1 70B Instruct, you can sample from Reflection Llama-3.1 70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
20
+
21
+ During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
22
+
23
+ This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
24
+
25
+ Inside the `<thinking>` section, the model may output one or more `<reflection>` tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
26
+
27
+ ## System Prompt
28
+
29
+ ```
30
+ The system prompt used for training this model is:
31
+
32
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
33
+
34
+ We recommend using this exact system prompt to get the best results from Reflection Llama-3.1 70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
35
+ ```
36
+
37
+ ## Chat Format
38
+
39
+ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:
40
+
41
+ ```
42
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
43
+
44
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
45
+
46
+ what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
47
+ ```
48
+
49
+ ## Tips for Performance
50
+
51
+ - We are initially recommending a `temperature` of `.7` and a `top_p` of `.95`.
52
+ - For increased accuracy, append `Think carefully.` at the end of your messages.
53
+
54
+ ## Dataset / Report
55
+
56
+ Both the dataset and a brief report detailing how we trained this model will be released next week, alongside our Reflection 405B model that we expect will be the top-performing LLM in the world, including closed-source models.
57
+
58
+ ---
59
+
60
+ Thanks to Jason Kuperberg and Josh Bickett from the [HyperWrite](https://hyperwriteai.com) team for reviewing drafts of the report we'll be releasing next week.
61
+
62
+ Also, we know right now the model is split into a ton of files. We'll condense this soon to make the model easier to download and work with!
Reflection-70B-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:505bd5179768edb96163e87bf7e9077c1875551a886a03e713f4a4c43feeec51
3
+ size 26375165344
Reflection-70B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b771f78127b382e3b91f3e88f7eb5683fe607ba0246191e7b5ce9867f3f0b0f6
3
+ size 37140654368
Reflection-70B-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3d03b32b9583269c4b19557bcb6a6b4a162e660d1a5250277275c740b542914
3
+ size 34267556128
Reflection-70B-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b627e282aebaf2455510a48d7934fbc48301eae4519b853ddb614d96a8378881
3
+ size 30912112928
Reflection-70B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8f2d33de48b6484d65dc3e81dc2790a17eac09576bfa6ba803da07a80a2cf63
3
+ size 42520461984
Reflection-70B-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:440b6fae4359e7eb00b4fcf3b185f6f7be0e5cdbe33ef070d2a4e7c20118e672
3
+ size 40347288224
Reflection-70B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a98a6b111c83709565aa23c0c393516f2f5fe6cfaa90619f5d8d33b4862777ae
3
+ size 49949891232
Reflection-70B-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a7a4f0b744a4b2f9a4cf432dff6e974636525d237954088a6d51cab8dd223de
3
+ size 48657521312