Upload folder using huggingface_hub
Browse files- README.md +66 -0
- config.json +80 -0
- model.safetensors +3 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +63 -0
- vocab.txt +0 -0
README.md
ADDED
@@ -0,0 +1,66 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language: tr
|
3 |
+
library_name: transformers
|
4 |
+
license: mit
|
5 |
+
tags:
|
6 |
+
- text-classification
|
7 |
+
- bert
|
8 |
+
- turkish
|
9 |
+
---
|
10 |
+
|
11 |
+
# BERT Din Kültürü Modeli (PyTorch)
|
12 |
+
Bu model, TensorFlow tabanlı `bert_din_kulturu_6` modelinin PyTorch formatına dönüştürülmüş versiyonudur.
|
13 |
+
|
14 |
+
## Model Performansı
|
15 |
+
Bu model, `dbmdz/bert-base-turkish-cased` modelinin `bert_din_kulturu_6` verisiyle eğitilmiş halidir.
|
16 |
+
Aşağıdaki sonuçları elde etmiştir:
|
17 |
+
|
18 |
+
- **Train Loss:** 0.2283
|
19 |
+
- **Train Accuracy:** 0.9475
|
20 |
+
- **Validation Loss:** 0.1557
|
21 |
+
- **Validation Accuracy:** 0.9614
|
22 |
+
- **Epochs:** 11
|
23 |
+
|
24 |
+
## Kullanım
|
25 |
+
Hugging Face `transformers` kütüphanesi ile kolayca kullanabilirsiniz:
|
26 |
+
```python
|
27 |
+
from transformers import AutoModelForSequenceClassification, AutoTokenizer
|
28 |
+
|
29 |
+
tokenizer = AutoTokenizer.from_pretrained("umutarpayy/bert_din_kulturu_6-pytorch")
|
30 |
+
model = AutoModelForSequenceClassification.from_pretrained("umutarpayy/bert_din_kulturu_6-pytorch")
|
31 |
+
```
|
32 |
+
|
33 |
+
## Eğitim Verisi
|
34 |
+
Bu model, özel bir veri seti ile eğitilmiştir. Detaylar aşağıdadır:
|
35 |
+
|
36 |
+
| Train Loss | Train Accuracy | Validation Loss | Validation Accuracy | Epoch |
|
37 |
+
|------------|---------------|----------------|----------------------|-------|
|
38 |
+
| 3.0220 | 0.2760 | 1.7334 | 0.5717 | 0 |
|
39 |
+
| 1.5144 | 0.5985 | 1.1455 | 0.6699 | 1 |
|
40 |
+
| 1.1069 | 0.6872 | 0.8943 | 0.7295 | 2 |
|
41 |
+
| 0.8911 | 0.7434 | 0.6902 | 0.7931 | 3 |
|
42 |
+
| 0.7160 | 0.7937 | 0.5481 | 0.8486 | 4 |
|
43 |
+
| 0.5892 | 0.8327 | 0.4402 | 0.8744 | 5 |
|
44 |
+
| 0.4838 | 0.8665 | 0.3478 | 0.9010 | 6 |
|
45 |
+
| 0.4034 | 0.8944 | 0.2653 | 0.9300 | 7 |
|
46 |
+
| 0.3421 | 0.9069 | 0.2125 | 0.9436 | 8 |
|
47 |
+
| 0.2890 | 0.9263 | 0.1814 | 0.9549 | 9 |
|
48 |
+
| 0.2487 | 0.9384 | 0.1629 | 0.9589 | 10 |
|
49 |
+
| 0.2283 | 0.9475 | 0.1557 | 0.9614 | 11 |
|
50 |
+
|
51 |
+
## Eğitim Yöntemi ve Parametreler
|
52 |
+
- **Optimizer:** Adam
|
53 |
+
- **Learning Rate:** 2e-5 (Polynomial Decay)
|
54 |
+
- **Beta1:** 0.9
|
55 |
+
- **Beta2:** 0.999
|
56 |
+
- **Epsilon:** 1e-8
|
57 |
+
- **Training Precision:** float32
|
58 |
+
|
59 |
+
## Framework Sürümleri
|
60 |
+
- **Transformers:** 4.47.1
|
61 |
+
- **TensorFlow:** 2.17.1
|
62 |
+
- **Datasets:** 3.2.0
|
63 |
+
- **Tokenizers:** 0.21.0
|
64 |
+
|
65 |
+
## Lisans
|
66 |
+
Bu model, MIT lisansına tabidir.
|
config.json
ADDED
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "umutarpayy/bert_din_kulturu_6",
|
3 |
+
"architectures": [
|
4 |
+
"BertForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"id2label": {
|
12 |
+
"0": "6.1.1. Peygamber ve peygamberlik kavramlar\u0131n\u0131 tan\u0131mlar.",
|
13 |
+
"1": "6.1.2. Peygamberlerin \u00f6zelliklerini ve g\u00f6revlerini a\u00e7\u0131klar.",
|
14 |
+
"2": "6.1.3. Peygamberlerde insanlar i\u00e7in g\u00fczel \u00f6rnekler oldu\u011funu fark eder.",
|
15 |
+
"3": "6.1.4. Vahyin g\u00f6nderili\u015f amac\u0131n\u0131 ara\u015ft\u0131r\u0131r.",
|
16 |
+
"4": "6.1.5. \u0130lahi kitaplar\u0131 ve g\u00f6nderildi\u011fi peygamberleri e\u015fle\u015ftirir.",
|
17 |
+
"5": "6.1.6. Hz. \u00c2dem\u2019in (a.s.) hayat\u0131n\u0131 ana hatlar\u0131yla \u00f6zetler.",
|
18 |
+
"6": "6.1.7. Kunut dualar\u0131n\u0131 okur, anlam\u0131n\u0131 s\u00f6yler.",
|
19 |
+
"7": "6.2.1. \u0130slam\u2019da namaz ibadetinin \u00f6nemini, ayet ve hadislerden \u00f6rneklerle a\u00e7\u0131klar.",
|
20 |
+
"8": "6.2.2. Namazlar\u0131, \u00e7e\u015fitlerine g\u00f6re s\u0131n\u0131fland\u0131r\u0131r.",
|
21 |
+
"9": "6.2.3. Namaz\u0131n k\u0131l\u0131n\u0131\u015f\u0131na \u00f6rnekler verir.",
|
22 |
+
"10": "6.2.4. Hz. Zekeriya\u2019n\u0131n (a.s.) hayat\u0131n\u0131 ana hatlar\u0131yla tan\u0131r.",
|
23 |
+
"11": "6.2.5. Fil suresini okur, anlam\u0131n\u0131 s\u00f6yler.",
|
24 |
+
"12": "6.3.1. \u0130slam dininin yasaklad\u0131\u011f\u0131 zararl\u0131 al\u0131\u015fkanl\u0131klara ayet ve hadislerden \u00f6rnekler verir.",
|
25 |
+
"13": "6.3.2. Zararl\u0131 al\u0131\u015fkanl\u0131klar\u0131n ba\u015flama sebeplerini sorgular.",
|
26 |
+
"14": "6.3.3. Zararl\u0131 al\u0131\u015fkanl\u0131klardan korunma yollar\u0131n\u0131 tart\u0131\u015f\u0131r.",
|
27 |
+
"15": "6.3.4. Zararl\u0131 al\u0131\u015fkanl\u0131klardan ka\u00e7\u0131nmaya istekli olur.",
|
28 |
+
"16": "6.3.5. Hz. Yahya\u2019n\u0131n (a.s.) hayat\u0131n\u0131 ana hatlar\u0131yla tan\u0131r.",
|
29 |
+
"17": "6.3.6. Tebbet suresini okur, anlam\u0131n\u0131 s\u00f6yler.",
|
30 |
+
"18": "6.4.1. Hz. Muhammed\u2019in (s.a.v.) davetinin Mekke D\u00f6nemini de\u011ferlendirir.",
|
31 |
+
"19": "6.4.2.\u00a0Medine\u2019ye hicretin sebep ve sonu\u00e7lar\u0131n\u0131 irdeler.",
|
32 |
+
"20": "6.4.3. Hz. Muhammed\u2019in (s.a.v.) davetinin Medine D\u00f6nemini de\u011ferlendirir.",
|
33 |
+
"21": "6.4.4. Nasr suresini okur, anlam\u0131n\u0131 s\u00f6yler.",
|
34 |
+
"22": "6.5.1. Toplumumuzu birle\u015ftiren temel de\u011ferleri fark eder.",
|
35 |
+
"23": "6.5.2. Din\u00ee bayramlar\u0131n ve \u00f6nemli g\u00fcn ve gecelerin toplumsal b\u00fct\u00fcnle\u015fmeye olan katk\u0131s\u0131n\u0131 yorumlar.",
|
36 |
+
"24": "6.5.4. Edebiyat\u0131m\u0131zdan dinin izlerine \u00f6rnekler bulur.",
|
37 |
+
"25": "6.5.5. \u00d6rf ve \u00e2detlerimizde yer alan din\u00ee unsurlar\u0131 fark eder."
|
38 |
+
},
|
39 |
+
"initializer_range": 0.02,
|
40 |
+
"intermediate_size": 3072,
|
41 |
+
"label2id": {
|
42 |
+
"6.1.1. Peygamber ve peygamberlik kavramlar\u0131n\u0131 tan\u0131mlar.": 0,
|
43 |
+
"6.1.2. Peygamberlerin \u00f6zelliklerini ve g\u00f6revlerini a\u00e7\u0131klar.": 1,
|
44 |
+
"6.1.3. Peygamberlerde insanlar i\u00e7in g\u00fczel \u00f6rnekler oldu\u011funu fark eder.": 2,
|
45 |
+
"6.1.4. Vahyin g\u00f6nderili\u015f amac\u0131n\u0131 ara\u015ft\u0131r\u0131r.": 3,
|
46 |
+
"6.1.5. \u0130lahi kitaplar\u0131 ve g\u00f6nderildi\u011fi peygamberleri e\u015fle\u015ftirir.": 4,
|
47 |
+
"6.1.6. Hz. \u00c2dem\u2019in (a.s.) hayat\u0131n\u0131 ana hatlar\u0131yla \u00f6zetler.": 5,
|
48 |
+
"6.1.7. Kunut dualar\u0131n\u0131 okur, anlam\u0131n\u0131 s\u00f6yler.": 6,
|
49 |
+
"6.2.1. \u0130slam\u2019da namaz ibadetinin \u00f6nemini, ayet ve hadislerden \u00f6rneklerle a\u00e7\u0131klar.": 7,
|
50 |
+
"6.2.2. Namazlar\u0131, \u00e7e\u015fitlerine g\u00f6re s\u0131n\u0131fland\u0131r\u0131r.": 8,
|
51 |
+
"6.2.3. Namaz\u0131n k\u0131l\u0131n\u0131\u015f\u0131na \u00f6rnekler verir.": 9,
|
52 |
+
"6.2.4. Hz. Zekeriya\u2019n\u0131n (a.s.) hayat\u0131n\u0131 ana hatlar\u0131yla tan\u0131r.": 10,
|
53 |
+
"6.2.5. Fil suresini okur, anlam\u0131n\u0131 s\u00f6yler.": 11,
|
54 |
+
"6.3.1. \u0130slam dininin yasaklad\u0131\u011f\u0131 zararl\u0131 al\u0131\u015fkanl\u0131klara ayet ve hadislerden \u00f6rnekler verir.": 12,
|
55 |
+
"6.3.2. Zararl\u0131 al\u0131\u015fkanl\u0131klar\u0131n ba\u015flama sebeplerini sorgular.": 13,
|
56 |
+
"6.3.3. Zararl\u0131 al\u0131\u015fkanl\u0131klardan korunma yollar\u0131n\u0131 tart\u0131\u015f\u0131r.": 14,
|
57 |
+
"6.3.4. Zararl\u0131 al\u0131\u015fkanl\u0131klardan ka\u00e7\u0131nmaya istekli olur.": 15,
|
58 |
+
"6.3.5. Hz. Yahya\u2019n\u0131n (a.s.) hayat\u0131n\u0131 ana hatlar\u0131yla tan\u0131r.": 16,
|
59 |
+
"6.3.6. Tebbet suresini okur, anlam\u0131n\u0131 s\u00f6yler.": 17,
|
60 |
+
"6.4.1. Hz. Muhammed\u2019in (s.a.v.) davetinin Mekke D\u00f6nemini de\u011ferlendirir.": 18,
|
61 |
+
"6.4.2.\u00a0Medine\u2019ye hicretin sebep ve sonu\u00e7lar\u0131n\u0131 irdeler.": 19,
|
62 |
+
"6.4.3. Hz. Muhammed\u2019in (s.a.v.) davetinin Medine D\u00f6nemini de\u011ferlendirir.": 20,
|
63 |
+
"6.4.4. Nasr suresini okur, anlam\u0131n\u0131 s\u00f6yler.": 21,
|
64 |
+
"6.5.1. Toplumumuzu birle\u015ftiren temel de\u011ferleri fark eder.": 22,
|
65 |
+
"6.5.2. Din\u00ee bayramlar\u0131n ve \u00f6nemli g\u00fcn ve gecelerin toplumsal b\u00fct\u00fcnle\u015fmeye olan katk\u0131s\u0131n\u0131 yorumlar.": 23,
|
66 |
+
"6.5.4. Edebiyat\u0131m\u0131zdan dinin izlerine \u00f6rnekler bulur.": 24,
|
67 |
+
"6.5.5. \u00d6rf ve \u00e2detlerimizde yer alan din\u00ee unsurlar\u0131 fark eder.": 25
|
68 |
+
},
|
69 |
+
"layer_norm_eps": 1e-12,
|
70 |
+
"max_position_embeddings": 512,
|
71 |
+
"model_type": "bert",
|
72 |
+
"num_attention_heads": 12,
|
73 |
+
"num_hidden_layers": 12,
|
74 |
+
"pad_token_id": 0,
|
75 |
+
"position_embedding_type": "absolute",
|
76 |
+
"transformers_version": "4.48.3",
|
77 |
+
"type_vocab_size": 2,
|
78 |
+
"use_cache": true,
|
79 |
+
"vocab_size": 32000
|
80 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:614b747578a91c1617a838f3220ffe4aaf6d559eb8ab7b574edb556bda0a3860
|
3 |
+
size 442572888
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4dd412615da4ff41ac303a941e9dc7afee37542d5f0cadf6248b02371c02090e
|
3 |
+
size 442615342
|
special_tokens_map.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"mask_token": {
|
10 |
+
"content": "[MASK]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "[PAD]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"sep_token": {
|
24 |
+
"content": "[SEP]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"unk_token": {
|
31 |
+
"content": "[UNK]",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
}
|
37 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"4": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": true,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_basic_tokenize": true,
|
47 |
+
"do_lower_case": false,
|
48 |
+
"extra_special_tokens": {},
|
49 |
+
"mask_token": "[MASK]",
|
50 |
+
"max_len": 512,
|
51 |
+
"max_length": 256,
|
52 |
+
"model_max_length": 512,
|
53 |
+
"never_split": null,
|
54 |
+
"pad_token": "[PAD]",
|
55 |
+
"sep_token": "[SEP]",
|
56 |
+
"stride": 0,
|
57 |
+
"strip_accents": null,
|
58 |
+
"tokenize_chinese_chars": true,
|
59 |
+
"tokenizer_class": "BertTokenizer",
|
60 |
+
"truncation_side": "right",
|
61 |
+
"truncation_strategy": "longest_first",
|
62 |
+
"unk_token": "[UNK]"
|
63 |
+
}
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|