--- language: - en license: other library_name: transformers tags: - chat - llama - finetune - peft base_model: duyhv1411/Llama-1.1B-qlora-ft model_name: TinyLlama-1.1B license_name: apache-2.0 license_link: https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md pipeline_tag: text-generation inference: false --- # duyhv1411/Llama-1.1B-qlora-ft This model is an advanced iteration of the powerful `TinyLlama/TinyLlama-1.1B-Chat-v1.0`, specifically fine-tuned to enhance its capabilities in generic domains. # ⚡ Quantized GGUF # How to use ```python # Use a pipeline as a high-level helper from transformers import pipeline prompt = """<|user|> Hello, how are you? <|assistant|> """ # Run our instruction-tuned model pipe = pipeline(task="text-generation", model="duyhv1411/Llama-1.1B-qlora-ft", return_full_text=False,) pipe(prompt)[0]["generated_text"] ```