File size: 788 Bytes
2b75344
 
 
 
 
d1ad19e
2b75344
 
e640043
2b75344
 
 
d1ad19e
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17





This is a merge of the Dolly LoRA with the main GPT-J-6B model, allowing users to use Dolly without having to worry about PEFT dependencies.


This hopes to be as similar as Alpaca, but without requirimg LLaMA access.

The performance is good but not as good as the orginal Alpaca trained from a base model of LLaMa

This is mostly due to the LLaMa 7B model being pretrained on 1T tokens and GPT-J-6B being trained on 300-400M tokens.
- LoRA originally trained by samwit, in: https://huggingface.co/samwit/dolly-lora
- The dataset is the cleaned version of the Alpaca dataset - https://github.com/gururise/AlpacaDataCleaned
- GPT-J-6b:  https://huggingface.co/EleutherAI/gpt-j-6B
- here is a Colab https://colab.research.google.com/drive/1O1JjyGaC300BgSJoUbru6LuWAzRzEqCz?usp=sharing