license: cc-by-nc-4.0 | |
language: | |
- en | |
tags: | |
- text-generation | |
datasets: | |
- stanford_alpaca | |
pipeline_tag: text-generation | |
This repo contains the full weights (8bit) for Falcon-7b | |
fit on the [Code Alpaca](https://huggingface.co/datasets/sahil2801/CodeAlpaca-20k) dataset. | |
## Reproduction | |
This version of the weights was trained with the following hyperparameters: | |
- Epochs: 6 | |
- Batch size: 128 | |
- Micro batch size: 8 | |
- Learning rate: 3e-4 | |
- Lora _r_: 16 | |
- Lora target modules: query_key_value | |
You can reproduce using this repository: | |
https://github.com/jina-ai/jerboa | |
Make sure you install requirements and finetune using this command using the following command: | |
``` | |
python finetune.py \ | |
--base-model='tiiuae/falcon-7b' \ | |
--num-epochs=6 \ | |
--output-dir='./jinaai/falcon-7b' \ | |
--lora-target-modules=query_key_value \ | |
--lora-r=16 \ | |
--micro-batch-size=8 | |
``` | |