|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- grammarly/coedit |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
tags: |
|
- torchtune |
|
- grammar-correction |
|
--- |
|
|
|
|
|
### Llama3 CoEdit |
|
|
|
This is a Llama3 8B based model trained using [torchtune](https://pytorch.org/torchtune) on the `grammarly/coedit` dataset. |
|
|
|
### Training details |
|
|
|
The exact training script (`lora_finetune_distributed`) and config (`8B_lora.yaml`) are both included in this repository. |
|
|
|
**Training command**: ``tune run --nproc_per_node 8 lora_finetune_distributed --config 8B_lora.yaml`` |
|
|
|
> Yes I used 8 GPUs :) |
|
|
|
In order to add the dataset, I added the following lines to the config: |
|
|
|
``` |
|
dataset: |
|
_component_: torchtune.datasets.instruct_dataset |
|
source: grammarly/coedit |
|
template: GrammarErrorCorrectionTemplate |
|
column_map: {"sentence": "src", "output": "tgt"} |
|
train_on_input: False |
|
split: train |
|
``` |
|
|
|
**Loss curve** |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/615b01ae487af9ad44dad803/Te9DycG2UVGm_JLnrG9De.png) |
|
|
|
|
|
### Evaluation results |