banghua commited on
Commit
2198427
1 Parent(s): 6d665b1

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -0
README.md ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ tags:
7
+ - RLHF
8
+ - Nexusflow
9
+ - Athene
10
+ - Chat Model
11
+ ---
12
+ # Athene-Llama3-70B
13
+
14
+ Blog: https://nexusflow.ai/blogs/introducing-athene-70b
15
+
16
+ <!-- Provide a quick summary of what the model is/does. -->
17
+
18
+ - **Developed by: The Nexusflow Team (Evan Frick\*, Peter Jin\*, Tianle Li\*, Karthik Ganesan, Jian Zhang, Jiantao Jiao and Banghua Zhu).
19
+ - **Model type:** Chat Model
20
+ - **Finetuned from model:** [Llama-3-70B-Instruct](meta-llama/Meta-Llama-3-70B-Instruct).
21
+
22
+
23
+ We introduce Athene-Llama3-70B, an open-weights LLM trained through RLHF based off Llama-3-70B-Instruct. Athene-70B achieves a high score on Arena-Hard-Auto, a proxy benchmark for Chatbot Arena.
24
+
25
+
26
+ | Model | Arena-Hard |
27
+ |-----------------------|------------------|
28
+ | GPT-4o (Proprietary) | 79.2% |
29
+ | Claude-3.5 (Proprietary) | 79.3% |
30
+ | **Athene-70B** (Open) | 77.8% |
31
+ | Gemini-Pro-1.5 (Proprietary) | 72.0% |
32
+ | Gemma-2-27B (Open) | 57.0% |
33
+ | Llama-3-70B (Open) | 46.6% |
34
+
35
+ ## Usage
36
+ Athene-70B uses the same chat template as Llama-3-70B-Instruct. Below is an example simple usage using the Transformers library.
37
+
38
+ ```Python
39
+ import transformers
40
+ import torch
41
+
42
+ model_id = "Nexusflow/Athene-70B"
43
+
44
+ pipeline = transformers.pipeline(
45
+ "text-generation",
46
+ model=model_id,
47
+ model_kwargs={"torch_dtype": torch.bfloat16},
48
+ device_map="auto",
49
+ )
50
+
51
+ messages = [
52
+ {"role": "system", "content": "You are an Athene Noctura, you can only speak with owl sounds. Whoooo whooo."},
53
+ {"role": "user", "content": "Whooo are you?"},
54
+ ]
55
+
56
+ terminators = [
57
+ pipeline.tokenizer.eos_token_id,
58
+ pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>")
59
+ ]
60
+
61
+ outputs = pipeline(
62
+ messages,
63
+ max_new_tokens=256,
64
+ eos_token_id=terminators,
65
+ do_sample=True,
66
+ temperature=0.6,
67
+ top_p=0.9,
68
+ )
69
+ print(outputs[0]["generated_text"][-1])
70
+
71
+ ```
72
+
73
+ ## Acknowledgment
74
+ We would like to thank the [LMSYS Organization](https://lmsys.org/) for their support of online demo and private test. We would like to thank Meta AI and the open source community for their efforts in providing the datasets and base models.
75
+
76
+ ## Citation
77
+ ```
78
+ @misc{Athene2024,
79
+ title = {Athene-70B: Redefining the Boundaries of Post-Training for Open Models},
80
+ url = {},
81
+ author = {Frick, Evan and Jin, Peter and Li, Tianle and Ganesan, Karthik and Zhang, Jian and Jiao, Jiantao and Zhu, Banghua},
82
+ month = {July},
83
+ year = {2024}
84
+ }
85
+ ```