jakeboggs commited on
Commit
53d1183
1 Parent(s): b779dd5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -144
README.md CHANGED
@@ -4,163 +4,48 @@ datasets:
4
  - jakeboggs/MTG-Eval
5
  ---
6
 
7
- # Model Card for Model ID
8
-
9
- MTG Llama is a QLoRA fine-tuned Llama 3 model on the [MTG Eval](https://huggingface.co/datasets/jakeboggs/MTG-Eval) dataset.
10
 
11
  ## Model Details
12
 
13
- ### Model Description
14
-
15
- <!-- Provide a longer summary of what this model is. -->
16
-
17
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
18
-
19
- - **Developed by:** Jake Boggs
20
- - **Model type:** [More Information Needed]
21
- - **Language(s) (NLP):** [More Information Needed]
22
- - **License:** [More Information Needed]
23
- - **Finetuned from model :** NousResearch/Meta-Llama-3-8B-Instruct
24
-
25
- ### Model Sources [optional]
26
-
27
- <!-- Provide the basic links for the model. -->
28
-
29
- - **Repository:** [More Information Needed]
30
- - **Paper [optional]:** [More Information Needed]
31
- - **Demo [optional]:** [More Information Needed]
32
-
33
- ## Uses
34
-
35
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
36
-
37
- ### Direct Use
38
-
39
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
40
-
41
- [More Information Needed]
42
-
43
- ### Downstream Use [optional]
44
-
45
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
46
-
47
- [More Information Needed]
48
-
49
- ### Out-of-Scope Use
50
-
51
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
52
-
53
- [More Information Needed]
54
-
55
- ## Bias, Risks, and Limitations
56
-
57
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
58
-
59
- [More Information Needed]
60
-
61
- ### Recommendations
62
-
63
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
64
-
65
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
66
-
67
- ## How to Get Started with the Model
68
-
69
- Use the code below to get started with the model.
70
-
71
- [More Information Needed]
72
-
73
- ## Training Details
74
-
75
- ### Training Data
76
-
77
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
78
-
79
- [More Information Needed]
80
-
81
- ### Training Procedure
82
-
83
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
84
-
85
- #### Preprocessing [optional]
86
-
87
- [More Information Needed]
88
-
89
-
90
- #### Training Hyperparameters
91
-
92
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
93
-
94
- #### Speeds, Sizes, Times [optional]
95
-
96
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
97
-
98
- [More Information Needed]
99
-
100
- ## Evaluation
101
-
102
- <!-- This section describes the evaluation protocols and provides the results. -->
103
-
104
- ### Testing Data, Factors & Metrics
105
-
106
- #### Testing Data
107
-
108
- <!-- This should link to a Dataset Card if possible. -->
109
-
110
- [More Information Needed]
111
-
112
- #### Factors
113
-
114
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
115
-
116
- [More Information Needed]
117
-
118
- #### Metrics
119
-
120
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
121
-
122
- [More Information Needed]
123
-
124
- ### Results
125
-
126
- [More Information Needed]
127
-
128
- #### Summary
129
-
130
-
131
-
132
- ## Model Examination [optional]
133
-
134
- <!-- Relevant interpretability work for the model goes here -->
135
-
136
- [More Information Needed]
137
-
138
- ## Technical Specifications [optional]
139
-
140
- ### Model Architecture and Objective
141
 
142
- [More Information Needed]
143
 
144
- ### Compute Infrastructure
145
 
146
- [More Information Needed]
147
 
148
- #### Hardware
 
 
 
149
 
150
- [More Information Needed]
151
 
152
- #### Software
 
 
 
153
 
154
- [More Information Needed]
155
 
156
- ## Citation [optional]
157
 
158
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
 
 
 
159
 
160
- **BibTeX:**
161
 
162
- [More Information Needed]
163
 
164
- **APA:**
165
 
166
- [More Information Needed]
 
 
4
  - jakeboggs/MTG-Eval
5
  ---
6
 
7
+ # Model Card for MTG-Llama: Fine-Tuned Model for Magic: The Gathering
 
 
8
 
9
  ## Model Details
10
 
11
+ **Model Name:** MTG-Llama
12
+ **Version:** 1.0
13
+ **Base Model:** Llama 3 8B Instruct
14
+ **Fine-Tuning Dataset:** MTG-Eval
15
+ **Author:** Jake Boggs
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
+ ## Model Description
18
 
19
+ MTG-Llama is a fine-tuned version of Llama 3 8B Instruct, tailored specifically for understanding and generating responses related to Magic: The Gathering (MTG). The model has been fine-tuned using a custom dataset, MTG-Eval, which includes question-answer pairs covering card descriptions, rules questions, and card interactions.
20
 
21
+ ## Intended Use
22
 
23
+ MTG-Llama is designed to assist users with:
24
+ - Generating deck construction ideas.
25
+ - Answering in-game rules questions.
26
+ - Understanding card interactions and abilities.
27
 
28
+ ## Training Data
29
 
30
+ The fine-tuning dataset, MTG-Eval, consists of 80,032 question-answer pairs generated synthetically. The dataset is categorized into:
31
+ - **Card Descriptions:** 26,702 examples
32
+ - **Rules Questions:** 27,104 examples
33
+ - **Card Interactions:** 26,226 examples
34
 
35
+ The data was sourced from the MTGJSON project and the Commander Spellbook combo database, reformatted into natural language question-answer pairs using ChatGPT 3.5.
36
 
37
+ ## Training Procedure
38
 
39
+ The model was fine-tuned using QLoRA with the following hyperparameters:
40
+ - **r:** 64
41
+ - **alpha:** 32
42
+ - **Steps:** 75
43
 
44
+ ## Acknowledgments
45
 
46
+ Thanks to the team at Commander Spellbook for generously sharing their dataset, without which this research would not be possible. All generated data is unofficial Fan Content permitted under the Fan Content Policy. Not approved/endorsed by Wizards. Portions of the materials used are property of Wizards of the Coast. ©Wizards of the Coast LLC.
47
 
48
+ ## Resources
49
 
50
+ - **Dataset:** [MTG-Eval on HuggingFace](https://huggingface.co/datasets/jakeboggs/MTG-Eval)
51
+ - **Training Code:** [GitHub Repository](https://github.com/JakeBoggs/Large-Language-Models-for-Magic-the-Gathering)