DeepMostInnovations commited on
Commit
7e8e702
·
verified ·
1 Parent(s): a3559ed

Upload Hindi embeddings model and all associated files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ hindi_corpus.txt filter=lfs diff=lfs merge=lfs -text
corpus_stats.json ADDED
The diff for this file is too large to render. See raw diff
 
embedding_model.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:343abe4c8b9a9e8f88fca1c5ab8bfc129c1ab2b3852c598e56ae7b3c3dedc1f1
3
- size 1491671832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:540a0bf4ad54049eb94815984cad54d3b569bab3a02317e1f0ee306d84f2e1d8
3
+ size 497221224
hindi_corpus.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2e1c771f83f7eca2fc1270939e14c12fd939b431be50308f07e52698111ba80
3
+ size 57984810
sample_corpus.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "unk_token": "<unk>",
5
+ "sep_token": "<sep>",
6
+ "pad_token": "<pad>",
7
+ "cls_token": "<cls>",
8
+ "mask_token": "<mask>"
9
+ }
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dcea7b417395ecf731062d1d241db13069e60e23ccc7cad4a77535900c38fe37
3
- size 1302086
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:123070be6221922fc42f22e54064f9cb2261715e02dd586a464231064d327d34
3
+ size 1290162
tokenizer.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,11 +1,33 @@
1
  {
2
- "model_type": "hindi_embedding",
3
- "tokenizer_class": "HindiSentencePieceTokenizer",
 
 
 
4
  "bos_token": "<s>",
5
  "eos_token": "</s>",
6
  "pad_token": "<pad>",
7
  "unk_token": "<unk>",
8
  "mask_token": "<mask>",
 
 
9
  "model_max_length": 512,
10
- "vocab_size": 50000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  }
 
1
  {
2
+ "name": "hindi-tokenizer",
3
+ "version": "1.0.0",
4
+ "model_type": "sentencepiece",
5
+ "sp_model_type": "unigram",
6
+ "tokenizer_class": "SentencePieceTokenizer",
7
  "bos_token": "<s>",
8
  "eos_token": "</s>",
9
  "pad_token": "<pad>",
10
  "unk_token": "<unk>",
11
  "mask_token": "<mask>",
12
+ "cls_token": "<cls>",
13
+ "sep_token": "<sep>",
14
  "model_max_length": 512,
15
+ "vocab_size": 50000,
16
+ "do_lower_case": false,
17
+ "special_tokens_map": {
18
+ "pad_token": "<pad>",
19
+ "unk_token": "<unk>",
20
+ "bos_token": "<s>",
21
+ "eos_token": "</s>",
22
+ "mask_token": "<mask>",
23
+ "sep_token": "<sep>",
24
+ "cls_token": "<cls>"
25
+ },
26
+ "tokenizer_file": "tokenizer.model",
27
+ "auto_map": {
28
+ "AutoTokenizer": [
29
+ "PreTrainedTokenizerFast",
30
+ null
31
+ ]
32
+ }
33
  }
training_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "hindi_tokenizer_sota",
3
+ "vocab_size": 50000,
4
+ "model_type": "unigram",
5
+ "character_coverage": 0.9995,
6
+ "num_iterations": 2000000,
7
+ "max_examples": 300000,
8
+ "sample_factor": 1.0,
9
+ "create_hf_config": true,
10
+ "debug": false,
11
+ "seed": 42
12
+ }