|
--- |
|
license: mit |
|
pipeline_tag: text2text-generation |
|
tags: |
|
- food |
|
datasets: |
|
- pocasrocas/recipe-gantt |
|
language: |
|
- en |
|
--- |
|
|
|
## Model Summary |
|
|
|
recipe-gantt is a monotask language model to translate recipes into gantt charts in a TSV format. |
|
|
|
It is a QLoRA finetune of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the [pocasrocas/recipe-gantt](https://huggingface.co/datasets/pocasrocas/recipe-gantt) dataset. I then used [llama.cpp](https://github.com/ggerganov/llama.cpp) to convert it to .gguf format for fast local inference. |
|
|
|
## Intended use |
|
|
|
To create recipe gantt charts and recipe gantt charts only π§βπ³ |
|
|
|
It's released here to accompany the [recipe-gantt tool](https://github.com/jbremz/recipe-gantt). |
|
|
|
## Input format |
|
|
|
The model was trained with the [alpaca](https://github.com/tatsu-lab/stanford_alpaca?tab=readme-ov-file#data-release) instruction/input/response prompt format. Exact details on the expected input can be inferred from the inference code [here](https://github.com/jbremz/recipe-gantt/blob/main/recipe-gantt.py). |
|
|
|
## Limitations |
|
|
|
- the model will inherit any limitations of the parent model (Mistral-7B-v0.1) |
|
- it was finetuned on a tiny synthetic dataset of only 288 examples so sometimes produces corrupted TSV files and populates cells inaccurately. |
|
|
|
## Training |
|
|
|
- QLoRA finetune using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) |
|
- ~1hr on NVIDIA GeForce RTX 3090 Ti ([wandb](https://wandb.ai/pocasrocas/recipe-gantt/runs/1ostj66y?workspace=user-pocasrocas)) |
|
- Training code [here](https://github.com/jbremz/recipe-gantt/tree/720a7e1b3594c03b46209671271fe3c2159fff16/notebooks/002-train) |