josh-oo commited on
Commit
43a4dcc
·
verified ·
1 Parent(s): 035c9d4

specter2_aspects_fold_0

Browse files
Files changed (2) hide show
  1. config.json +64 -17
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,29 +1,76 @@
1
  {
2
- "_name_or_path": "BAAI/bge-m3",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "architectures": [
4
- "AspectModelXLMRoberta"
 
 
 
 
 
 
5
  ],
6
- "aspects": [],
7
  "attention_probs_dropout_prob": 0.1,
8
- "bos_token_id": 0,
9
  "classifier_dropout": null,
10
- "eos_token_id": 2,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
- "hidden_size": 1024,
14
  "initializer_range": 0.02,
15
- "intermediate_size": 4096,
16
- "layer_norm_eps": 1e-05,
17
- "max_position_embeddings": 8194,
18
- "model_type": "xlm-roberta",
19
- "num_attention_heads": 16,
20
- "num_hidden_layers": 24,
21
- "output_past": true,
22
- "pad_token_id": 1,
23
  "position_embedding_type": "absolute",
24
- "torch_dtype": "bfloat16",
 
25
  "transformers_version": "4.48.3",
26
- "type_vocab_size": 1,
27
  "use_cache": true,
28
- "vocab_size": 250002
29
  }
 
1
  {
2
+ "_name_or_path": "specter2",
3
+ "adapters": {
4
+ "adapters": {
5
+ "specter2": "9076f36a74755ac4"
6
+ },
7
+ "config_map": {
8
+ "9076f36a74755ac4": {
9
+ "adapter_residual_before_ln": false,
10
+ "cross_adapter": false,
11
+ "dropout": 0.0,
12
+ "factorized_phm_W": true,
13
+ "factorized_phm_rule": false,
14
+ "hypercomplex_nonlinearity": "glorot-uniform",
15
+ "init_weights": "bert",
16
+ "inv_adapter": null,
17
+ "inv_adapter_reduction_factor": null,
18
+ "is_parallel": false,
19
+ "learn_phm": true,
20
+ "leave_out": [],
21
+ "ln_after": false,
22
+ "ln_before": false,
23
+ "mh_adapter": false,
24
+ "non_linearity": "relu",
25
+ "original_ln_after": true,
26
+ "original_ln_before": true,
27
+ "output_adapter": true,
28
+ "phm_bias": true,
29
+ "phm_c_init": "normal",
30
+ "phm_dim": 4,
31
+ "phm_init_range": 0.0001,
32
+ "phm_layer": false,
33
+ "phm_rank": 1,
34
+ "reduction_factor": 16,
35
+ "residual_before_ln": true,
36
+ "scaling": 1.0,
37
+ "shared_W_phm": false,
38
+ "shared_phm_rule": true,
39
+ "stochastic_depth": 0.0,
40
+ "use_gating": false
41
+ }
42
+ },
43
+ "fusion_config_map": {},
44
+ "fusion_name_map": {},
45
+ "fusions": {}
46
+ },
47
  "architectures": [
48
+ "AspectModelBert"
49
+ ],
50
+ "aspects": [
51
+ "participants",
52
+ "intervention",
53
+ "condition",
54
+ "outcome"
55
  ],
 
56
  "attention_probs_dropout_prob": 0.1,
 
57
  "classifier_dropout": null,
 
58
  "hidden_act": "gelu",
59
  "hidden_dropout_prob": 0.1,
60
+ "hidden_size": 768,
61
  "initializer_range": 0.02,
62
+ "intermediate_size": 3072,
63
+ "layer_norm_eps": 1e-12,
64
+ "max_position_embeddings": 512,
65
+ "model_type": "bert",
66
+ "num_attention_heads": 12,
67
+ "num_hidden_layers": 12,
68
+ "pad_token_id": 0,
 
69
  "position_embedding_type": "absolute",
70
+ "prediction_heads": {},
71
+ "torch_dtype": "float32",
72
  "transformers_version": "4.48.3",
73
+ "type_vocab_size": 2,
74
  "use_cache": true,
75
+ "vocab_size": 31094
76
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6aa6333c76f5921dbd2a2f591fdb81191cc5866424ff8d9d263b76f6c8f0729a
3
- size 1135554922
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5afa48b65fe1c156defaa29d6ba6b5915474458f6490d06eea1c088772af3d3
3
+ size 439708708