Existance commited on
Commit
a2e651a
·
verified ·
1 Parent(s): cddea52

Existance/CIS_qlora_summarization_model

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: peft
3
- license: apache-2.0
4
- base_model: Qwen/Qwen2.5-0.5B
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,9 +14,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # qlora_summarization
16
 
17
- This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B](https://huggingface.co/Qwen/Qwen2.5-0.5B) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 8.3791
20
 
21
  ## Model description
22
 
@@ -39,7 +39,7 @@ The following hyperparameters were used during training:
39
  - train_batch_size: 2
40
  - eval_batch_size: 2
41
  - seed: 42
42
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 3
45
  - mixed_precision_training: Native AMP
@@ -48,15 +48,15 @@ The following hyperparameters were used during training:
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
- | 8.4191 | 1.0 | 2025 | 8.4432 |
52
- | 8.5048 | 2.0 | 4050 | 8.3952 |
53
- | 8.3704 | 3.0 | 6075 | 8.3791 |
54
 
55
 
56
  ### Framework versions
57
 
58
  - PEFT 0.14.0
59
- - Transformers 4.47.0
60
- - Pytorch 2.5.1+cu121
61
- - Datasets 3.3.1
62
- - Tokenizers 0.21.0
 
1
  ---
2
  library_name: peft
3
+ license: llama3.2
4
+ base_model: meta-llama/Llama-3.2-3B
5
  tags:
6
  - generated_from_trainer
7
  model-index:
 
14
 
15
  # qlora_summarization
16
 
17
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 7.7306
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 2
40
  - eval_batch_size: 2
41
  - seed: 42
42
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 3
45
  - mixed_precision_training: Native AMP
 
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
+ | 7.6614 | 1.0 | 2025 | 7.7951 |
52
+ | 7.7233 | 2.0 | 4050 | 7.7536 |
53
+ | 7.4209 | 3.0 | 6075 | 7.7306 |
54
 
55
 
56
  ### Framework versions
57
 
58
  - PEFT 0.14.0
59
+ - Transformers 4.50.0
60
+ - Pytorch 2.6.0+cu124
61
+ - Datasets 3.5.0
62
+ - Tokenizers 0.21.1