--- license: apache-2.0 language: - en - ja tags: - finetuned library_name: transformers pipeline_tag: text-generation --- # Model Card for Ninja-v1-128k The Mistral-7B--based Large Language Model (LLM) is an noveldataset fine-tuned version of the Mistral-7B-v0.1 VecTeus has the following changes compared to Mistral-7B-v0.1. - 128k context window (8k context in v0.1) - Achieving both high quality Japanese and English generation - Memory ability that does not forget even after long-context generation This model was created with the help of GPUs from the first LocalAI hackathon. We would like to take this opportunity to thank ## List of Creation Methods - Chatvector for multiple models - Simple linear merging of result models - Domain and Sentence Enhancement with LORA - Context expansion ## Instruction format Freed from templates. Congratulations ## Example prompts to improve (Japanese) - BAD: あなたは○○として振る舞います - GOOD: あなたは○○です - BAD: あなたは○○ができます - GOOD: あなたは○○をします # Other points to keep in mind If possible, we recommend inferring with llamacpp rather than Transformers.