hz3519 commited on
Commit
09b8939
1 Parent(s): d0b3b6e

Add models

Browse files
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - "en"
4
+ thumbnail: "https://example.com/path/to/your/thumbnail.jpg" # URL to a thumbnail used in social sharing
5
+ tags:
6
+ - "tag1" # For example, "sentiment-analysis"
7
+ - "tag2" # For example, "machine-translation"
8
+ license: "mit"
9
+ datasets:
10
+ - "dataset1" # For example, "imdb"
11
+ - "dataset2" # For example, "wmt16"
12
+ metrics:
13
+ - "metric1" # For example, "accuracy"
14
+ - "metric2" # For example, "f1"
15
+ ---
16
+
17
+ # Your Model Name
18
+
19
+ ## Introduction
20
+
21
+ This is a brief introduction about your transformer-based model. Here, you can mention the type of the model, the task it was trained for, its performance, and other key features or highlights.
22
+
23
+ ## Training
24
+
25
+ Here, give detailed information about how the model was trained:
26
+
27
+ - Dataset(s) used for training
28
+ - Preprocessing techniques used
29
+ - Training configuration such as the batch size, learning rate, optimizer, number of epochs, etc.
30
+ - Any specific challenges or notable aspects of the training process
31
+
32
+ ## Usage
33
+
34
+ Provide examples of how to use the model for inference. You can provide both a simple usage case and a more complex one if necessary. Make sure to explain what the inputs and outputs are.
35
+
36
+ Here's a basic example:
37
+
38
+ from transformers import AutoTokenizer, AutoModel
39
+
40
+ tokenizer = AutoTokenizer.from_pretrained("your-model-name")
41
+ model = AutoModel.from_pretrained("your-model-name")
42
+
43
+ inputs = tokenizer("Your example sentence", return_tensors="pt")
44
+ outputs = model(**inputs)
45
+
46
+ # Explain what the outputs are
47
+
48
+ ## Evaluation
49
+
50
+ Discuss how the model was evaluated, which metrics were used, and what results it achieved.
51
+
52
+ ## Limitations and Bias
53
+
54
+ Every model has its limitations and may have certain biases due to the data it was trained on. Explain those here.
55
+
56
+ ## About Us
57
+
58
+ A small introduction about you or your team.
59
+
60
+ ## Acknowledgments
61
+
62
+ Thank people, organizations or mention the resources that helped you in this work.
63
+
64
+ ## License
65
+
66
+ This model is distributed under the MIT license.
67
+
68
+ ## Contact
69
+
70
+ Provide a contact method (e.g., email or GitHub issues) for people to reach out with questions, comments, or concerns.
71
+
72
+ ## References
73
+
74
+ List any relevant references for your model here.
.ipynb_checkpoints/model_evaluate-checkpoint.ipynb ADDED
The diff for this file is too large to render. See raw diff
 
model_M/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 24, "query_size": 512, "key_size": 512, "value_size": 512, "num_hiddens": 512, "num_layers": 6, "dropout": 0.2, "lr": 0.0004, "training_steps": 300000, "batch_size": 4096, "label_smoothing": 0.1, "ffn_num_input": 512, "ffn_num_hiddens": 2048, "num_heads": 8, "norm_shape": [512], "device": "cpu"}
model_M/model_weights.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:197a52e0a47711602b2ae60f2f9ee34b15fdc20b195aad66ab8df0bda9aa3256
3
+ size 176577875
model_M_retrain/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 24, "query_size": 512, "key_size": 512, "value_size": 512, "num_hiddens": 512, "num_layers": 6, "dropout": 0.2, "lr": 0.0004, "training_steps": 300000, "batch_size": 4096, "label_smoothing": 0.1, "ffn_num_input": 512, "ffn_num_hiddens": 2048, "num_heads": 8, "norm_shape": [512], "device": "cpu"}
model_M_retrain/model_weights.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:596171bfa552fd963cf7e3f1bd8b6c6f35f7ea14d129d9f7b451d0ecb4014ddc
3
+ size 176577875