Pretergeek commited on
Commit
6b0c5ca
·
verified ·
1 Parent(s): f7d288c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -1
README.md CHANGED
@@ -5,4 +5,14 @@ datasets:
5
  base_model:
6
  - openchat/openchat-3.5-0106
7
  library_name: transformers
8
- ---
 
 
 
 
 
 
 
 
 
 
 
5
  base_model:
6
  - openchat/openchat-3.5-0106
7
  library_name: transformers
8
+ ---
9
+ <p align="center">
10
+ <a href="https://ko-fi.com/pretergeek">Buy me a Ko-Fi</a> •
11
+ <a href="https://patreon.com/Pretergeek">Support my work using Patreon</a>
12
+ </p>
13
+
14
+ # OpenChat-3.5-0106_32K-PoSE
15
+
16
+ This model is [Openchat-3.5-0106](https://huggingface.co/openchat/openchat-3.5-0106) with the context length extended from 8192 tokens to 32768 tokens using [PoSE](https://huggingface.co/papers/2309.10400).
17
+
18
+ The model was fine-tuned using [Rank-Stabilized LoRA](https://huggingface.co/blog/damjan-k/rslora) and the [LongAlpaca-12K](Yukang/LongAlpaca-12k) dataset. I hope to continue extending the context in future versions and then apply the same methods to my [upscaled versions of OpenChat-3.5](https://huggingface.co/collections/Pretergeek/openchat-35-0106-with-additional-layers-66a8d3262c7c3ebdd7783a29).