rwitz2 commited on
Commit
1e785d5
·
1 Parent(s): 9cda601

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +108 -0
README.md ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ ---
7
+ Merge:
8
+ ```
9
+ slices:
10
+ - sources:
11
+ - model: viethq188/LeoScorpius-7B-Chat-DPO
12
+ layer_range: [0, 32]
13
+ - model: GreenNode/GreenNodeLM-7B-v1olet
14
+ layer_range: [0, 32]
15
+ merge_method: slerp
16
+ base_model: viethq188/LeoScorpius-7B-Chat-DPO
17
+ parameters:
18
+ t:
19
+ - filter: self_attn
20
+ value: [0, 0.5, 0.3, 0.7, 1]
21
+ - filter: mlp
22
+ value: [1, 0.5, 0.7, 0.3, 0]
23
+ - value: 0.5 # fallback for rest of tensors
24
+ dtype: float16
25
+ ```
26
+
27
+
28
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63a259d0f30c46422789d38d/tmdM1fjNAmzV125zWd3_J.png)
29
+
30
+ # Go Bruins V2.1 - A Fine-tuned Language Model
31
+
32
+ ## Updates
33
+
34
+ ## Overview
35
+
36
+ **Go Bruins-V2** is a language model fine-tuned on the rwitz/go-bruins architecture. It's designed to push the boundaries of NLP applications, offering unparalleled performance in generating human-like text.
37
+
38
+ ## Model Details
39
+
40
+ - **Developer:** Ryan Witzman
41
+ - **Base Model:** [rwitz/go-bruins](https://huggingface.co/rwitz/go-bruins)
42
+ - **Fine-tuning Method:** Direct Preference Optimization (DPO)
43
+ - **Training Steps:** 642
44
+ - **Language:** English
45
+ - **License:** MIT
46
+
47
+ ## Capabilities
48
+
49
+ Go Bruins excels in a variety of NLP tasks, including but not limited to:
50
+ - Text generation
51
+ - Language understanding
52
+ - Sentiment analysis
53
+
54
+ ## Usage
55
+
56
+ **Warning:** This model may output NSFW or illegal content. Use with caution and at your own risk.
57
+
58
+ ### For Direct Use:
59
+
60
+ ```python
61
+ from transformers import pipeline
62
+
63
+ model_name = "rwitz/go-bruins-v2"
64
+ inference_pipeline = pipeline('text-generation', model=model_name)
65
+
66
+ input_text = "Your input text goes here"
67
+ output = inference_pipeline(input_text)
68
+
69
+ print(output)
70
+ ```
71
+
72
+
73
+ ### Not Recommended For:
74
+
75
+ - Illegal activities
76
+ - Harassment
77
+ - Professional advice or crisis situations
78
+
79
+ ## Training and Evaluation
80
+
81
+ Trained on a dataset from [athirdpath/DPO_Pairs-Roleplay-Alpaca-NSFW](https://huggingface.co/datasets/athirdpath/DPO_Pairs-Roleplay-Alpaca-NSFW), Go Bruins V2 has shown promising improvements over its predecessor, Go Bruins.
82
+
83
+ # Evaluations
84
+
85
+ | Metric | Average | Arc Challenge | Hella Swag | MMLU | Truthful Q&A | Winogrande | GSM8k |
86
+ |---------------|---------|---------------|------------|------|--------------|------------|-------|
87
+ | **Score** | 72.07 | 69.8 | 87.05| 64.75 | 59.7 | 81.45 | 69.67 |
88
+
89
+ Note: The original MMLU evaluation has been corrected to include 5-shot data rather than 1-shot data.
90
+ ## Contact
91
+
92
+ For any inquiries or feedback, reach out to Ryan Witzman on Discord: `rwitz_`.
93
+
94
+ ---
95
+ ## Citations
96
+ ```
97
+ @misc{unacybertron7b,
98
+ title={Cybertron: Uniform Neural Alignment},
99
+ author={Xavier Murias},
100
+ year={2023},
101
+ publisher = {HuggingFace},
102
+ journal = {HuggingFace repository},
103
+ howpublished = {\url{https://huggingface.co/fblgit/una-cybertron-7b-v2-bf16}},
104
+ }
105
+ ```
106
+
107
+ *This model card was created with care by Ryan Witzman.*
108
+ rewrite this model card for new version called go-bruins-v2 that is finetuned on dpo on the original go-bruins model on athirdpath/DPO_Pairs-Roleplay-Alpaca-NSFW