Leon-Leee commited on
Commit
3b3be01
·
verified ·
1 Parent(s): 8de40b2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md CHANGED
@@ -1,3 +1,63 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - Leon-Leee/wizardlm_evol_instruct_v2_196K_backuped
5
+ - m-a-p/Code-Feedback
6
+ - openbmb/UltraInteract_sft
7
+ - ise-uiuc/Magicoder-Evol-Instruct-110K
8
+ language:
9
+ - en
10
+ metrics:
11
+ - code_eval
12
+ library_name: transformers
13
+ tags:
14
+ - code
15
  ---
16
+ ## AIGCodeGeek-DS-6.7B
17
+
18
+ ### Introduction
19
+ AIGCodeGeek-DS-6.7B is the first version of our Code-LLM family with competitive performance on benchmarks such as HumanEval(+) and MBPP(+).
20
+ It gains a lot of insights from the open-source community and we deeply appreciate all of these great works.
21
+ We are preparing for the tech report, so stay tuned for more details.
22
+
23
+ ### Model Details
24
+ #### Model Description
25
+ - Developed by: [Leon Li](https://huggingface.co/Leon-Leee)
26
+ - License: [DeepSeek](https://github.com/deepseek-ai/DeepSeek-Coder/blob/main/LICENSE-MODEL)
27
+ - Fine-tuned from [deepseek-ai/deepseek-coder-6.7b-base](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-base) with full parameters
28
+
29
+ ### Training data
30
+ A mixture of both
31
+ - samples from several high-quality open-source datasets (read *Acknowledgements*),
32
+ - our private datasets (already decontaminated with benchmarks).
33
+
34
+ ### Evaluation
35
+ To check out our evaluation results: [EvalPlus](https://evalplus.github.io/leaderboard.html)
36
+
37
+ ### Requirements
38
+ It should work with the same requirements as DeepSeek-Coder-6.7B
39
+ ```torch>=2.0
40
+ tokenizers>=0.14.0
41
+ transformers>=4.35.0
42
+ accelerate
43
+ sympy>=1.12
44
+ pebble
45
+ timeout-decorator
46
+ attrdict
47
+ ```
48
+
49
+
50
+ ### QuickStart
51
+ TBD
52
+ ```
53
+ ```
54
+
55
+ ### Limits
56
+
57
+
58
+ ### Acknowledgements
59
+ - [WizardCoder](https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder): WizardLM-Evol-Instruct V2 datasets
60
+ - We used a back-up([Leon-Leee/wizardlm_evol_instruct_v2_196K_backuped](https://huggingface.co/datasets/Leon-Leee/wizardlm_evol_instruct_v2_196K_backuped)) since this dataset has been deleted.
61
+ - [Magicoder](https://github.com/ise-uiuc/magicoder/): [Magicoder-Evol-Instruct-110K](https://huggingface.co/datasets/ise-uiuc/Magicoder-Evol-Instruct-110K) from theblackcat102/evol-codealpaca-v1(https://huggingface.co/datasets/theblackcat102/evol-codealpaca-v1)
62
+ - [Eurus](https://github.com/OpenBMB/Eurus): reasoning enhancement dataset of [openbmb/UltraInteract_sft](https://huggingface.co/datasets/openbmb/UltraInteract_sft)
63
+ - [OpenCoderInterpreter](https://opencodeinterpreter.github.io/): [m-a-p/Code-Feedback](https://huggingface.co/datasets/m-a-p/Code-Feedback)