File size: 1,550 Bytes
c8faae9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
license: other
license_name: qwen
license_link: >-
  https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT
language:
- en
- zh
tags:
- qwen
- qwen1.5
- qwen2
- llama
inference: false
---
## Description
This repo containst the "LLaMAfied" version of [Qwen1.5-72B-Chat](https://huggingface.co/Qwen/Qwen1.5-72B-Chat) by Alibaba Cloud. I used the amazing [script](https://github.com/Minami-su/character_AI_open/blob/main/llamafy_qwen_v2.py) made by [Minami-su](https://huggingface.co/Minami-su) to LLaMAfy the model.

## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
tokenizer = AutoTokenizer.from_pretrained("sayhan/Qwen1.5-72B-Chat-LLaMAfied")
model = AutoModelForCausalLM.from_pretrained("sayhan/Qwen1.5-72B-Chat-LLaMAfied", torch_dtype="auto", device_map="auto")
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)

messages = [
    {"role": "user", "content": "Who are you?"}
]
inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
inputs = inputs.to("cuda")
generate_ids = model.generate(inputs,max_length=2048, streamer=streamer)
```

## Other LLaMAfied Qwen1.5 Models
The two other sizes of the Qwen1.5 have been LLaMAfied by [Minami-su](https://huggingface.co/Minami-su)
- **0.5B:** [Minami-su/Qwen1.5-0.5B-Chat_llamafy](https://huggingface.co/Minami-su/Qwen1.5-0.5B-Chat_llamafy)
- **7B:** [Minami-su/Qwen1.5-7B-Chat_llamafy](https://huggingface.co/Minami-su/Qwen1.5-7B-Chat_llamafy)