--- library_name: transformers license: apache-2.0 language: - fr - en datasets: - jpacifico/French-Alpaca-dataset-Instruct-110K --- ## Model Card for Model ID A 7B language model. Good in French. ![image/jpeg](https://github.com/jpacifico/French-Alpaca/blob/main/Assets/French-Alpaca_500px.png?raw=true) ### Model Description The French-Alpaca is a 7.24B params LLM model based on the Mistral-7B-Instruct-v0.2 foundation model, fine-tuned from the original French-Alpaca-dataset entirely generated with OpenAI GPT-3.5-turbo. French-Alpaca is a general model and can itself be finetuned to be specialized for specific use cases. The fine-tuning method is inspired from https://crfm.stanford.edu/2023/03/13/alpaca.html ### Limitations The French-Alpaca model is a quick demonstration that a base 7B model can be easily fine-tuned to specialize in a particular language. It does not have any moderation mechanisms. - **Developed by:** Jonathan Pacifico, 2024 - **Model type:** LLM - **Language(s) (NLP):** French - **License:** Apache 2.0 - **Finetuned from model:** mistralai/Mistral-7B-Instruct-v0.2