|
--- |
|
datasets: |
|
- jondurbin/airoboros-gpt4-1.4.1 |
|
--- |
|
|
|
NOTE: This LoRA was trained on Llama-30b AFTER additional pretraining. I intend on providing the LoRA of that pretraining too. Applying this LoRA to base Llama-30b will likely result in a performance reduction. I have uploaded the fp16 merged weights [here](https://huggingface.co/bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-LoRA/) |
|
|
|
Mostly untested! |
|
|
|
Find GPTQ quantized weights and full model card here: https://huggingface.co/bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-GPTQ |
|
|
|
# RoPE Scaled QLoRA Fine-tune of Llama-33b on airoboros-gpt4-1.4.1 (LoRA) |
|
|
|
## Overview |
|
|
|
This is [Jon Durbin's Airoboros 33B GPT4 1.4](https://huggingface.co/jondurbin/airoboros-33b-gpt4-1.4) (LoRA) with several key modifications: |
|
- Context length extended to 16384 by RoPE Scaled Embeddings. |
|
- The Llama-33b base model is pretrained for additional 100 steps on 8192 length sequences from the pile dataset. |
|
- Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4 |
|
|
|
**This is a QLoRA fine-tune** |
|
|
|
Pretraining took 10 hours. Finetuning took ~41 hours on 1x RTX 6000 Ada. |