AtifAli121 commited on
Commit
6c66b40
·
verified ·
1 Parent(s): 9194292

End of training

Browse files
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # sqlcoder-7b-sqli
18
 
19
- This model is a fine-tuned version of [defog/sqlcoder-7b](https://huggingface.co/defog/sqlcoder-7b) on the None dataset.
20
 
21
  ## Model description
22
 
@@ -39,12 +39,12 @@ The following hyperparameters were used during training:
39
  - train_batch_size: 1
40
  - eval_batch_size: 8
41
  - seed: 42
42
- - gradient_accumulation_steps: 8
43
- - total_train_batch_size: 8
44
  - optimizer: Use OptimizerNames.PAGED_ADAMW with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
- - lr_scheduler_warmup_ratio: 0.03
47
- - training_steps: 200
48
 
49
  ### Training results
50
 
@@ -55,5 +55,5 @@ The following hyperparameters were used during training:
55
  - PEFT 0.15.2.dev0
56
  - Transformers 4.51.3
57
  - Pytorch 2.6.0+cu124
58
- - Datasets 3.5.0
59
  - Tokenizers 0.21.1
 
16
 
17
  # sqlcoder-7b-sqli
18
 
19
+ This model is a fine-tuned version of [defog/sqlcoder-7b](https://huggingface.co/defog/sqlcoder-7b) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 1
40
  - eval_batch_size: 8
41
  - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 4
44
  - optimizer: Use OptimizerNames.PAGED_ADAMW with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_ratio: 0.05
47
+ - training_steps: 100
48
 
49
  ### Training results
50
 
 
55
  - PEFT 0.15.2.dev0
56
  - Transformers 4.51.3
57
  - Pytorch 2.6.0+cu124
58
+ - Datasets 3.6.0
59
  - Tokenizers 0.21.1
adapter_config.json CHANGED
@@ -13,7 +13,7 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 64,
17
  "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
@@ -22,16 +22,16 @@
22
  "lm_head"
23
  ],
24
  "peft_type": "LORA",
25
- "r": 32,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
 
29
  "up_proj",
30
- "q_proj",
31
  "k_proj",
32
- "v_proj",
33
- "gate_proj",
34
- "o_proj"
35
  ],
36
  "task_type": "CAUSAL_LM",
37
  "trainable_token_indices": null,
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 32,
17
  "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
 
22
  "lm_head"
23
  ],
24
  "peft_type": "LORA",
25
+ "r": 16,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "gate_proj",
30
+ "v_proj",
31
  "up_proj",
 
32
  "k_proj",
33
+ "o_proj",
34
+ "q_proj"
 
35
  ],
36
  "task_type": "CAUSAL_LM",
37
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e450e66eae402e0645fdd5854970f01c1eb3389d4249310250e0fe004ef7968
3
- size 392219992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef11d96bdff5e4087096bb77a64550becce58d96d3a284a8fdd55bd5db3f11ae
3
+ size 327208280
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3a5363549a83cfc81d3adb05ec1814b146f207f83b737b16393b872db380f18
3
- size 4518449388
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9c2db11f29c8c063ea9726100cd651a31379213f88e877aa2390f132d375434
3
+ size 4453437644
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 384,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e170a7de62521457c6d18349b9f1cb02a4a7bed3b41640e677241635b28a213e
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45b6e0cd9a15580cf37d3bfa5fd0ef557593f1eeca984b4a1041b538d99824b6
3
  size 5304