Update README.md
Browse files
README.md
CHANGED
@@ -6,6 +6,8 @@ datasets:
|
|
6 |
|
7 |
## Flan-UL2-Alpaca
|
8 |
|
|
|
|
|
9 |
This [Github repository](https://github.com/ConiferLabsWA/flan-ul2-alpaca) contains code for leveraging the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) synthetic dataset to fine tune the [Flan-UL2](https://huggingface.co/google/flan-ul2) model, leveraging recent advances in instruction tuning. The Flan UL2 model has been shown to outperform Flan-T5 XXL on a number of metrics and has a 4x improvement in receptive field (2048 vs 512).
|
10 |
|
11 |
### Resource Considerations
|
|
|
6 |
|
7 |
## Flan-UL2-Alpaca
|
8 |
|
9 |
+
Model weights are from epoch 0.
|
10 |
+
|
11 |
This [Github repository](https://github.com/ConiferLabsWA/flan-ul2-alpaca) contains code for leveraging the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) synthetic dataset to fine tune the [Flan-UL2](https://huggingface.co/google/flan-ul2) model, leveraging recent advances in instruction tuning. The Flan UL2 model has been shown to outperform Flan-T5 XXL on a number of metrics and has a 4x improvement in receptive field (2048 vs 512).
|
12 |
|
13 |
### Resource Considerations
|