Update README.md
Browse files
README.md
CHANGED
@@ -2,6 +2,10 @@ This is the first model which converts Qwen2.5-7B's checkpoint to RWKV-7 archite
|
|
2 |
It's trained in one server with 8xA800 for one day which might not be that versatile. It shows an acceptable performance to chat with you fluently.
|
3 |
The shortage is that this base model can't do math and related tasks. I'll add a more balanced data to improve that model's capability later.
|
4 |
|
5 |
-
Please refer the https://github.com/yynil/RWKVinLLAMA/blob/rwkv_7/
|
6 |
|
|
|
|
|
|
|
|
|
7 |
|
|
|
2 |
It's trained in one server with 8xA800 for one day which might not be that versatile. It shows an acceptable performance to chat with you fluently.
|
3 |
The shortage is that this base model can't do math and related tasks. I'll add a more balanced data to improve that model's capability later.
|
4 |
|
5 |
+
Please refer the https://github.com/yynil/RWKVinLLAMA/blob/rwkv_7/tests/test_chat_cli.py about how to use it.
|
6 |
|
7 |
+
```python
|
8 |
+
python tests/test_chat_cli.py --config_file configs/qwen_7b.yaml --is_hybrid --num_gpus 1
|
9 |
+
```
|
10 |
+
You may need to change the configuration yaml to point the path to the checkpoint path you downloaded.
|
11 |
|