acecalisto3 commited on
Commit
d78c69d
·
verified ·
1 Parent(s): 73a041b

Create config.json

Browse files
Files changed (1) hide show
  1. config.json +72 -0
config.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AceCalisto3"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 1024,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 4096,
13
+ "layer_norm_eps": 1e-05,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "acecalisto3",
16
+ "num_attention_heads": 16,
17
+ "num_hidden_layers": 24,
18
+ "pad_token_id": 1,
19
+ "vocab_size": 50257,
20
+ "use_cache": true,
21
+ "gradient_checkpointing": false,
22
+ "is_decoder": true,
23
+ "is_encoder_decoder": false,
24
+ "tie_word_embeddings": false,
25
+ "use_bias": true,
26
+ "vocab_path": "vocab.json",
27
+ "use_bf16": false,
28
+ "use_fp16": false,
29
+ "use_m32": false,
30
+ "use_m64": false,
31
+ "use_m128": false,
32
+ "use_sharded_weights": false,
33
+ "use_mps": false,
34
+ "use_xpu": false,
35
+ "use_tpu": false,
36
+ "use_fast_gelu": false,
37
+ "use_gelu_new": false,
38
+ "torchscript": false,
39
+ "load_in_8bit": false,
40
+ "load_in_4bit": false,
41
+ "quantization_config": {
42
+ "bits": 8,
43
+ "mode": "static",
44
+ "use_fp16": false
45
+ },
46
+ "trust_remote_code": false,
47
+ "use_slow_tokenizer": false,
48
+ "use_auth_token": "YOUR_HUGGING_FACE_TOKEN",
49
+ "revision": "main",
50
+ "cache_dir": "./model_cache",
51
+ "resume_download": true,
52
+ "proxies": null,
53
+ "local_files_only": false,
54
+ "force_download": false,
55
+ "token": null,
56
+ "repo_id": "acecalisto3/instructiphi",
57
+ "repo_type": "model",
58
+ "use_cdn": true,
59
+ "force_http": false,
60
+ "max_retries": 5,
61
+ "timeout": 60,
62
+ "use_auth_token": null,
63
+ "ignore_download_warnings": false,
64
+ "offline_mode": false,
65
+ "use_fast_tokenizer": true,
66
+ "use_fast": true,
67
+ "low_cpu": false,
68
+ "pad_to_max_length": false,
69
+ "model_max_length": 512,
70
+ "max_position_embeddings": 512,
71
+ "model_kwargs": {}
72
+ }