naot97 commited on
Commit
8e97f59
·
1 Parent(s): 7d2547c

Upload BloomForCausalLM

Browse files
Files changed (3) hide show
  1. config.json +22 -15
  2. generation_config.json +1 -1
  3. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,26 +1,33 @@
1
  {
2
- "_name_or_path": "ura-hcmut/ura-llama-7b",
 
3
  "architectures": [
4
- "LlamaForCausalLM"
5
  ],
 
 
 
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
8
- "hidden_act": "silu",
9
- "hidden_size": 4096,
10
  "initializer_range": 0.02,
11
- "intermediate_size": 11008,
12
- "max_position_embeddings": 4096,
13
- "model_type": "llama",
14
- "num_attention_heads": 32,
15
- "num_hidden_layers": 32,
16
- "num_key_value_heads": 32,
17
- "pad_token_id": 0,
 
18
  "pretraining_tp": 1,
19
- "rms_norm_eps": 1e-06,
20
- "rope_scaling": null,
21
- "tie_word_embeddings": false,
 
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.31.0",
 
24
  "use_cache": true,
25
- "vocab_size": 32000
26
  }
 
1
  {
2
+ "_name_or_path": "bigscience/bloomz-560m",
3
+ "apply_residual_connection_post_layernorm": false,
4
  "architectures": [
5
+ "BloomForCausalLM"
6
  ],
7
+ "attention_dropout": 0.0,
8
+ "attention_softmax_in_fp32": true,
9
+ "bias_dropout_fusion": true,
10
  "bos_token_id": 1,
11
  "eos_token_id": 2,
12
+ "hidden_dropout": 0.0,
13
+ "hidden_size": 1024,
14
  "initializer_range": 0.02,
15
+ "layer_norm_epsilon": 1e-05,
16
+ "masked_softmax_fusion": true,
17
+ "model_type": "bloom",
18
+ "n_head": 16,
19
+ "n_inner": null,
20
+ "n_layer": 24,
21
+ "offset_alibi": 100,
22
+ "pad_token_id": 3,
23
  "pretraining_tp": 1,
24
+ "seq_length": 2048,
25
+ "skip_bias_add": true,
26
+ "skip_bias_add_qkv": false,
27
+ "slow_but_exact": false,
28
  "torch_dtype": "bfloat16",
29
  "transformers_version": "4.31.0",
30
+ "unk_token_id": 0,
31
  "use_cache": true,
32
+ "vocab_size": 250880
33
  }
generation_config.json CHANGED
@@ -2,6 +2,6 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "pad_token_id": 0,
6
  "transformers_version": "4.31.0"
7
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "pad_token_id": 3,
6
  "transformers_version": "4.31.0"
7
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:461c054294b716079c27230f800784948c421890cad18b9a7ecfb54e507bf298
3
- size 7952798001
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:000a459797c7b64a7927994d26cfc037b5fab8aa18ecaeac277073e6a3c94157
3
+ size 1118528353