CHZY-1 commited on
Commit
5dc9814
1 Parent(s): d09968e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -11
README.md CHANGED
@@ -5,18 +5,21 @@ license: cc-by-sa-4.0
5
  tags:
6
  - trl
7
  - sft
8
- - generated_from_trainer
 
 
 
9
  model-index:
10
  - name: sqlcoder-7b-2_FineTuned_PEFT_QLORA_adapter
11
  results: []
 
 
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
 
17
- # sqlcoder-7b-2_FineTuned_PEFT_QLORA_adapter
18
 
19
- This model is a fine-tuned version of [defog/sqlcoder-7b-2](https://huggingface.co/defog/sqlcoder-7b-2) on an unknown dataset.
20
 
21
  ## Model description
22
 
@@ -24,11 +27,15 @@ More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
28
 
29
  ## Training and evaluation data
30
 
31
- More information needed
 
 
 
 
32
 
33
  ## Training procedure
34
 
@@ -47,10 +54,6 @@ The following hyperparameters were used during training:
47
  - num_epochs: 5
48
  - mixed_precision_training: Native AMP
49
 
50
- ### Training results
51
-
52
-
53
-
54
  ### Framework versions
55
 
56
  - PEFT 0.13.2
 
5
  tags:
6
  - trl
7
  - sft
8
+ - QLora
9
+ - peft
10
+ - SQL
11
+ - causal-lm
12
  model-index:
13
  - name: sqlcoder-7b-2_FineTuned_PEFT_QLORA_adapter
14
  results: []
15
+ language:
16
+ - en
17
  ---
18
 
 
 
19
 
20
+ # sqlcoder-7b-2_FineTuned_QLORA_Adapter
21
 
22
+ This model is a fine-tuned version of [defog/sqlcoder-7b-2](https://huggingface.co/defog/sqlcoder-7b-2) on 260 MS SQL examples (Task, Schema and Answer pairs).
23
 
24
  ## Model description
25
 
 
27
 
28
  ## Intended uses & limitations
29
 
30
+ SQL Query Generation
31
 
32
  ## Training and evaluation data
33
 
34
+ This model was re-trained using the QLoRA method, with the following configurations:
35
+ - 'r = 32', 'lora_alpha = 32', 'lora_dropout = 0.1', bias='none', task_type='CAUSAL_LM'
36
+
37
+ Quantization parameters:
38
+ - load_in_4bit=True, bnb_4bit_quant_type="nf4", bnb_4bit_compute_dtype=torch.bfloat16
39
 
40
  ## Training procedure
41
 
 
54
  - num_epochs: 5
55
  - mixed_precision_training: Native AMP
56
 
 
 
 
 
57
  ### Framework versions
58
 
59
  - PEFT 0.13.2