recipe-gantt-v0.1 / README.md
pocasrocas's picture
Update README.md
8ce2375 verified
---
license: mit
pipeline_tag: text2text-generation
tags:
- food
datasets:
- pocasrocas/recipe-gantt
language:
- en
---
## Model Summary
recipe-gantt is a monotask language model to translate recipes into gantt charts in a TSV format.
It is a QLoRA finetune of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the [pocasrocas/recipe-gantt](https://huggingface.co/datasets/pocasrocas/recipe-gantt) dataset. I then used [llama.cpp](https://github.com/ggerganov/llama.cpp) to convert it to .gguf format for fast local inference.
## Intended use
To create recipe gantt charts and recipe gantt charts only πŸ§‘β€πŸ³
It's released here to accompany the [recipe-gantt tool](https://github.com/jbremz/recipe-gantt).
## Input format
The model was trained with the [alpaca](https://github.com/tatsu-lab/stanford_alpaca?tab=readme-ov-file#data-release) instruction/input/response prompt format. Exact details on the expected input can be inferred from the inference code [here](https://github.com/jbremz/recipe-gantt/blob/main/recipe-gantt.py).
## Limitations
- the model will inherit any limitations of the parent model (Mistral-7B-v0.1)
- it was finetuned on a tiny synthetic dataset of only 288 examples so sometimes produces corrupted TSV files and populates cells inaccurately.
## Training
- QLoRA finetune using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl)
- ~1hr on NVIDIA GeForce RTX 3090 Ti ([wandb](https://wandb.ai/pocasrocas/recipe-gantt/runs/1ostj66y?workspace=user-pocasrocas))
- Training code [here](https://github.com/jbremz/recipe-gantt/tree/720a7e1b3594c03b46209671271fe3c2159fff16/notebooks/002-train)