deuswoof commited on
Commit
4a49bec
·
1 Parent(s): 3d74971

Training in progress, step 10

Browse files
26_10_23_config_test.csv CHANGED
@@ -6,7 +6,7 @@ run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop
6
  5,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
7
  6,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
8
  7,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
9
- 8,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
10
  9,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
11
  10,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
12
  11,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
 
6
  5,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
7
  6,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
8
  7,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
9
+ 8,classification_of_valuems set True,True,2,100,0.8,False,True,False,False
10
  9,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
11
  10,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
12
  11,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
26_10_23_results_test.csv CHANGED
@@ -26,4 +26,4 @@ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnes
26
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
27
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
28
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
29
- 0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
 
26
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
27
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
28
  0,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
29
+ 0,1528.8057931953265,classification_of_valuems set True,17.268380403518677,0,0.2478655270791332,0.3576253534106945,0.2778231048355419,0.2832970005606974,0.4013383893240502,0.3126046416067341,0.3252384765751978,0.4466121889283891,0.3509020400982625,0.1004556235113589,0.1415746397318885,0.1108019406524126,0.1231052130296416,0.1705496820565623,0.1340090789111166,0.1477750597812323,0.19929804844641,0.1574531105220511,0.25,0.25,0.25
27_10_23_23_results_real_3.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
+ 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
27_10_23_config_test.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop_token,classification_of_valuems,stemming,lemmatization
2
+ 1,no variations,False,2,100,0.8,False,False,False,False
3
+ 2,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
27_10_23_results_real_3.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
+ 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
27_10_23_results_test.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
+ 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
README.md CHANGED
@@ -1000,6 +1000,18 @@ The following `bitsandbytes` quantization config was used during training:
1000
  - bnb_4bit_use_double_quant: True
1001
  - bnb_4bit_compute_dtype: bfloat16
1002
 
 
 
 
 
 
 
 
 
 
 
 
 
1003
  The following `bitsandbytes` quantization config was used during training:
1004
  - quant_method: bitsandbytes
1005
  - load_in_8bit: False
@@ -1096,5 +1108,6 @@ The following `bitsandbytes` quantization config was used during training:
1096
  - PEFT 0.5.0
1097
  - PEFT 0.5.0
1098
  - PEFT 0.5.0
 
1099
 
1100
  - PEFT 0.5.0
 
1000
  - bnb_4bit_use_double_quant: True
1001
  - bnb_4bit_compute_dtype: bfloat16
1002
 
1003
+ The following `bitsandbytes` quantization config was used during training:
1004
+ - quant_method: bitsandbytes
1005
+ - load_in_8bit: False
1006
+ - load_in_4bit: True
1007
+ - llm_int8_threshold: 6.0
1008
+ - llm_int8_skip_modules: None
1009
+ - llm_int8_enable_fp32_cpu_offload: False
1010
+ - llm_int8_has_fp16_weight: False
1011
+ - bnb_4bit_quant_type: nf4
1012
+ - bnb_4bit_use_double_quant: True
1013
+ - bnb_4bit_compute_dtype: bfloat16
1014
+
1015
  The following `bitsandbytes` quantization config was used during training:
1016
  - quant_method: bitsandbytes
1017
  - load_in_8bit: False
 
1108
  - PEFT 0.5.0
1109
  - PEFT 0.5.0
1110
  - PEFT 0.5.0
1111
+ - PEFT 0.5.0
1112
 
1113
  - PEFT 0.5.0
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44b0e4641a21400f9ac8b892d5ea86ba119fb8bfd1e2ae0c92a38bef210714a7
3
  size 100733709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79395823541aa379c4f344f6d050486319a8431f84a8bdd7e1a0d50f6dfc715d
3
  size 100733709
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc30f8a1f8cbc7e6e739a3bccd24e9771a9a66560696bd18ec550294cffdb65d
3
- size 261131840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea8d177806c82c8311bd985e84d7ca45c95ac8df57ddadbd9bfa3a9503194ada
3
+ size 100690288
config.json CHANGED
@@ -1,33 +1,33 @@
1
  {
2
- "_name_or_path": "tiiuae/falcon-7b",
3
- "alibi": false,
4
  "apply_residual_connection_post_layernorm": false,
5
  "architectures": [
6
  "FalconForCausalLM"
7
  ],
8
  "attention_dropout": 0.0,
9
  "auto_map": {
10
- "AutoConfig": "tiiuae/falcon-7b--configuration_falcon.FalconConfig",
11
- "AutoModel": "tiiuae/falcon-7b--modeling_falcon.FalconModel",
12
- "AutoModelForCausalLM": "tiiuae/falcon-7b--modeling_falcon.FalconForCausalLM",
13
- "AutoModelForQuestionAnswering": "tiiuae/falcon-7b--modeling_falcon.FalconForQuestionAnswering",
14
- "AutoModelForSequenceClassification": "tiiuae/falcon-7b--modeling_falcon.FalconForSequenceClassification",
15
- "AutoModelForTokenClassification": "tiiuae/falcon-7b--modeling_falcon.FalconForTokenClassification"
16
  },
17
- "bias": false,
18
- "bos_token_id": 11,
19
- "eos_token_id": 11,
20
  "hidden_dropout": 0.0,
21
- "hidden_size": 4544,
22
  "initializer_range": 0.02,
23
  "layer_norm_epsilon": 1e-05,
24
  "model_type": "falcon",
25
- "multi_query": true,
26
  "new_decoder_architecture": false,
27
- "num_attention_heads": 71,
28
- "num_hidden_layers": 32,
29
- "num_kv_heads": 71,
30
- "parallel_attn": true,
31
  "quantization_config": {
32
  "bnb_4bit_compute_dtype": "bfloat16",
33
  "bnb_4bit_quant_type": "nf4",
@@ -43,5 +43,5 @@
43
  "torch_dtype": "bfloat16",
44
  "transformers_version": "4.34.1",
45
  "use_cache": false,
46
- "vocab_size": 65024
47
  }
 
1
  {
2
+ "_name_or_path": "tiiuae/falcon-rw-1b",
3
+ "alibi": true,
4
  "apply_residual_connection_post_layernorm": false,
5
  "architectures": [
6
  "FalconForCausalLM"
7
  ],
8
  "attention_dropout": 0.0,
9
  "auto_map": {
10
+ "AutoConfig": "tiiuae/falcon-rw-1b--configuration_falcon.FalconConfig",
11
+ "AutoModel": "tiiuae/falcon-rw-1b--modeling_falcon.FalconModel",
12
+ "AutoModelForCausalLM": "tiiuae/falcon-rw-1b--modeling_falcon.FalconForCausalLM",
13
+ "AutoModelForQuestionAnswering": "tiiuae/falcon-rw-1b--modeling_falcon.FalconForQuestionAnswering",
14
+ "AutoModelForSequenceClassification": "tiiuae/falcon-rw-1b--modeling_falcon.FalconForSequenceClassification",
15
+ "AutoModelForTokenClassification": "tiiuae/falcon-rw-1b--modeling_falcon.FalconForTokenClassification"
16
  },
17
+ "bias": true,
18
+ "bos_token_id": 1,
19
+ "eos_token_id": 2,
20
  "hidden_dropout": 0.0,
21
+ "hidden_size": 2048,
22
  "initializer_range": 0.02,
23
  "layer_norm_epsilon": 1e-05,
24
  "model_type": "falcon",
25
+ "multi_query": false,
26
  "new_decoder_architecture": false,
27
+ "num_attention_heads": 32,
28
+ "num_hidden_layers": 24,
29
+ "num_kv_heads": 32,
30
+ "parallel_attn": false,
31
  "quantization_config": {
32
  "bnb_4bit_compute_dtype": "bfloat16",
33
  "bnb_4bit_quant_type": "nf4",
 
43
  "torch_dtype": "bfloat16",
44
  "transformers_version": "4.34.1",
45
  "use_cache": false,
46
+ "vocab_size": 50304
47
  }
proc_dataset.csv CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:748da944795546b277c6e7e61d431e15d33d32dc5c1ea3807fa94ebde545837d
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2be2e6c88b186be5fb357782826030044d7a59da0f0613012d4cb7645606a2fe
3
  size 4283