VatsaDev commited on
Commit
d041c8a
·
1 Parent(s): a127c77

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +84 -1
README.md CHANGED
@@ -1,3 +1,86 @@
1
  ---
2
  license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
+ ---
4
+
5
+ # nanoChatGpt
6
+ a barebones Nanogpt, but finetuned on conversational data
7
+
8
+ [Colab link](https://colab.research.google.com/drive/1a2aW5eClKjHVQJp-qtHDz4m6ai4yh49Z?usp=sharing)
9
+ all updates in [updates.md](updates.md)
10
+ ## how does it work?
11
+
12
+ This is a fork of Nanogpt, but trained on the data format of a chatbot like chatgpt, with the format inspired by `oasst-pythia-12b`
13
+
14
+ ```
15
+ <human> ... <endOfText>
16
+ <Bot> ... <endOfText>
17
+ <human> ... <endOfText>
18
+ <Bot> ... <endOfText>
19
+ <human> ... <endOfText>
20
+ <Bot> ... <endOfText>
21
+ ```
22
+
23
+ ## Demo
24
+ to view its capabilites, simply run the colab, the whole thing should take around ~30 minutes on a t4 instance. If you look at both outputs, Chat has no prompting, but directing the chat in a direction is very helpful
25
+
26
+ ## Limitations
27
+
28
+ I did not make the data dumps/corpuses that make up this data, and can't account for any biases, as the dataset it self is based off the conversations of real people who may or may not have had biases. The model is meant for academic research purposes, and isn't meant for any important or high risk scenarios. Do not follow its advice
29
+
30
+ ### whats in the data
31
+ - personaChat corpus
32
+ - twitter Dumps corpus
33
+ - Ubuntu dialog corpus
34
+ - Chatbot arena conversations (not for commercial use)
35
+ - ParlAI empathetic Chat (not for commercial use)
36
+ - Hackernews Dumps corpus
37
+ - BabyLM challenge speech data (Unsure, but probably not for commercial use)
38
+ - Eleuthier AI arithmetic dataset
39
+ - Oasst-1
40
+
41
+ for commercial purposes, just take the files `input1.txt` through `input36.txt`
42
+
43
+ ## Features
44
+
45
+ * Medium Dataset(~630mb), full of a variety of conversations, and a little arithmetic
46
+ * can talk to you on a variety of topics, smoothly switch between topics, and often sounds like a real person
47
+ * GPT-2-medium 353 million parameters
48
+ * Very Fast Inference on GPU
49
+ * User to Bot chat
50
+ * chat.py has functions similar to openai api stop, removes all content after a certain word
51
+ * Option to use google bard to look through and improve conversations
52
+
53
+ ## Problems / TODOs
54
+
55
+ * ### Math and Logical Reasoning
56
+
57
+ * While mostly out of the scope of this model, this is something for future models, the two main options are to detect math and run backend code, or train the model with Chain-of-though math prompting, which is probably more feasible, For logical reasoning, I might try incorporating datasets like `garage-bAInd/Open-Platypus`, etc
58
+
59
+ * ### no memory/recall
60
+
61
+ * Though probably out of scope for this model, this is something for future models, With many models, you can ask what you were just talking about or to summarize the conversation above. When that is attempted with this model:
62
+
63
+ ```plaintext
64
+ Human: Dogecoin is cool
65
+ Bot: indeed, very shibe
66
+ Human: what were we just talking about?
67
+ Bot: me and a friend gave up on mining, but now I can
68
+ ```
69
+
70
+ as we can see, it continues on with a sentence on mining, confirming that it understood the context(GPT2 info) but it does not have the ability to recall. I suspect that has to do with the models data, and that if I were to feed it data like short-context calling and summarization data, it would gain those abilities
71
+
72
+
73
+ <br /><br /><br />
74
+ Anyone Who can contribute to the repo, please do so, any and all contributions are welcome, simply add a little to the dataset and expand it dataset would be amazing.
75
+
76
+ #### citations
77
+ ```
78
+ @misc{zheng2023judging,
79
+ title={Judging LLM-as-a-judge with MT-Bench and Chatbot Arena},
80
+ author={Lianmin Zheng and Wei-Lin Chiang and Ying Sheng and Siyuan Zhuang and Zhanghao Wu and Yonghao Zhuang and Zi Lin and Zhuohan Li and Dacheng Li and Eric. P Xing and Hao Zhang and Joseph E. Gonzalez and Ion Stoica},
81
+ year={2023},
82
+ eprint={2306.05685},
83
+ archivePrefix={arXiv},
84
+ primaryClass={cs.CL}
85
+ }
86
+ ```