rai-sandeep commited on
Commit
40aa8c1
·
1 Parent(s): ae6b854

Delete my-trained-model-4

Browse files
my-trained-model-4/README.md DELETED
@@ -1,32 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
-
18
- The following `bitsandbytes` quantization config was used during training:
19
- - load_in_8bit: False
20
- - load_in_4bit: True
21
- - llm_int8_threshold: 6.0
22
- - llm_int8_skip_modules: None
23
- - llm_int8_enable_fp32_cpu_offload: False
24
- - llm_int8_has_fp16_weight: False
25
- - bnb_4bit_quant_type: nf4
26
- - bnb_4bit_use_double_quant: True
27
- - bnb_4bit_compute_dtype: bfloat16
28
- ### Framework versions
29
-
30
- - PEFT 0.4.0.dev0
31
-
32
- - PEFT 0.4.0.dev0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my-trained-model-4/adapter_config.json DELETED
@@ -1,19 +0,0 @@
1
- {
2
- "base_model_name_or_path": "vilsonrodrigues/falcon-7b-instruct-sharded",
3
- "bias": "none",
4
- "fan_in_fan_out": false,
5
- "inference_mode": true,
6
- "init_lora_weights": true,
7
- "layers_pattern": null,
8
- "layers_to_transform": null,
9
- "lora_alpha": 32,
10
- "lora_dropout": 0.05,
11
- "modules_to_save": null,
12
- "peft_type": "LORA",
13
- "r": 16,
14
- "revision": null,
15
- "target_modules": [
16
- "query_key_value"
17
- ],
18
- "task_type": "CAUSAL_LM"
19
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my-trained-model-4/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:22ce225184f52f7b2083b1f6b35ed63ed9ffe0ec73ae71ccee69ca83886ffc13
3
- size 18898161
 
 
 
 
my-trained-model-4/special_tokens_map.json DELETED
@@ -1,17 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- ">>TITLE<<",
4
- ">>ABSTRACT<<",
5
- ">>INTRODUCTION<<",
6
- ">>SUMMARY<<",
7
- ">>COMMENT<<",
8
- ">>ANSWER<<",
9
- ">>QUESTION<<",
10
- ">>DOMAIN<<",
11
- ">>PREFIX<<",
12
- ">>SUFFIX<<",
13
- ">>MIDDLE<<"
14
- ],
15
- "eos_token": "<|endoftext|>",
16
- "pad_token": "<|endoftext|>"
17
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my-trained-model-4/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
my-trained-model-4/tokenizer_config.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "clean_up_tokenization_spaces": true,
4
- "eos_token": "<|endoftext|>",
5
- "model_max_length": 2048,
6
- "tokenizer_class": "PreTrainedTokenizerFast"
7
- }