davidsi commited on
Commit
d57bab1
·
verified ·
1 Parent(s): 3ddf860

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -6
README.md CHANGED
@@ -1,13 +1,15 @@
1
  ---
2
  library_name: transformers
3
- tags: []
 
 
4
  ---
5
 
6
  # Model Card for Model ID
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
-
11
 
12
  ## Model Details
13
 
@@ -17,13 +19,13 @@ tags: []
17
 
18
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
 
20
- - **Developed by:** [More Information Needed]
21
  - **Funded by [optional]:** [More Information Needed]
22
  - **Shared by [optional]:** [More Information Needed]
23
  - **Model type:** [More Information Needed]
24
  - **Language(s) (NLP):** [More Information Needed]
25
  - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
 
28
  ### Model Sources [optional]
29
 
@@ -69,12 +71,22 @@ Users (both direct and downstream) should be made aware of the risks, biases and
69
 
70
  ## How to Get Started with the Model
71
 
72
- Use the code below to get started with the model.
 
 
 
 
 
 
 
73
 
74
  [More Information Needed]
75
 
76
  ## Training Details
77
 
 
 
 
78
  ### Training Data
79
 
80
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
@@ -154,7 +166,8 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
154
 
155
  ### Model Architecture and Objective
156
 
157
- [More Information Needed]
 
158
 
159
  ### Compute Infrastructure
160
 
 
1
  ---
2
  library_name: transformers
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
  ---
7
 
8
  # Model Card for Model ID
9
 
10
  <!-- Provide a quick summary of what the model is/does. -->
11
 
12
+ Model finetuned specifically for expertise on AMD technologies and python coding.
13
 
14
  ## Model Details
15
 
 
19
 
20
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
21
 
22
+ - **Developed by:** David Silverstein
23
  - **Funded by [optional]:** [More Information Needed]
24
  - **Shared by [optional]:** [More Information Needed]
25
  - **Model type:** [More Information Needed]
26
  - **Language(s) (NLP):** [More Information Needed]
27
  - **License:** [More Information Needed]
28
+ - **Finetuned from model meta-llama/Meta-Llama-3.1-8B-Instruct**
29
 
30
  ### Model Sources [optional]
31
 
 
71
 
72
  ## How to Get Started with the Model
73
 
74
+ Use the code below to get started with the model.
75
+
76
+ import torch
77
+ from transformers import AutoTokenizer, AutoModelForCausalLM
78
+
79
+ model_name = 'davidsi/Llama3_1-8B-Instruct-AMD-python'
80
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
81
+ llm = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16)
82
 
83
  [More Information Needed]
84
 
85
  ## Training Details
86
 
87
+ Torchtune was used for full finetuning, for 5 epochs on a single Instinct MI210 GPU. The training set consisted
88
+ of 1658 question/answer pairs in Alpaca format.
89
+
90
  ### Training Data
91
 
92
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
 
166
 
167
  ### Model Architecture and Objective
168
 
169
+ This model is a finetuned version of Llama 3.1, which is an auto-regressive language model that uses
170
+ an optimized transformer architecture.
171
 
172
  ### Compute Infrastructure
173