xwinxu commited on
Commit
1917b5d
1 Parent(s): c2e45b9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +43 -0
README.md ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - stanfordnlp/SHP
5
+ - Anthropic/hh-rlhf
6
+ - OpenAssistant/oasst1
7
+ language:
8
+ - en
9
+ metrics:
10
+ - accuracy
11
+ tags:
12
+ - human feedback
13
+ - rlhf
14
+ - preferences
15
+ - alignment
16
+ - HALO
17
+ - halos
18
+ - dpo
19
+ - rl
20
+ ---
21
+
22
+ ![halos](https://gist.github.com/assets/29318529/fe2d8391-dbd1-4b7e-9dc4-7cb97e55bc06)
23
+
24
+ This repo contains the model checkpoints for:
25
+ - model family <b>pythia1-4b</b>
26
+ - optimized with the loss <b>SFT+CSFT</b>
27
+ - aligned using the SHP, Anthropic HH and Open Assistant datasets.
28
+
29
+ To prompt archangel models, ensure that the format is consistent with that of TuluV2, i.e. `"<s>\n<|user|>\n" + <prompt> + "\n<|assistant|>\n</s>"`.
30
+ Note that the BOS / EOS tokens should be excluded if automatically added by your tokenizer during batch collation.
31
+
32
+ Please refer to our [code repository](https://github.com/ContextualAI/HALOs) or [blog](https://contextual.ai/better-cheaper-faster-llm-alignment-with-kto/) which contains intructions for training your own HALOs and links to our model cards.
33
+
34
+ If you find this repo or the technical paper useful in your research, please feel free to cite [our work](https://github.com/ContextualAI/HALOs/blob/main/assets/report.pdf):
35
+ ```
36
+ @techreport{ethayarajh2023halos,
37
+ author = {Ethayarajh, Kawin and Xu, Winnie, and Jurafsky, Dan and Kiela, Douwe},
38
+ title = {Human-Centered Loss Functions (HALOs)},
39
+ institution = {Contextual AI},
40
+ note = {https://github.com/ContextualAI/HALOs/blob/main/assets/report.pdf},
41
+ year = {2023},
42
+ }
43
+ ```