Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,63 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- Leon-Leee/wizardlm_evol_instruct_v2_196K_backuped
|
5 |
+
- m-a-p/Code-Feedback
|
6 |
+
- openbmb/UltraInteract_sft
|
7 |
+
- ise-uiuc/Magicoder-Evol-Instruct-110K
|
8 |
+
language:
|
9 |
+
- en
|
10 |
+
metrics:
|
11 |
+
- code_eval
|
12 |
+
library_name: transformers
|
13 |
+
tags:
|
14 |
+
- code
|
15 |
---
|
16 |
+
## AIGCodeGeek-DS-6.7B
|
17 |
+
|
18 |
+
### Introduction
|
19 |
+
AIGCodeGeek-DS-6.7B is the first version of our Code-LLM family with competitive performance on benchmarks such as HumanEval(+) and MBPP(+).
|
20 |
+
It gains a lot of insights from the open-source community and we deeply appreciate all of these great works.
|
21 |
+
We are preparing for the tech report, so stay tuned for more details.
|
22 |
+
|
23 |
+
### Model Details
|
24 |
+
#### Model Description
|
25 |
+
- Developed by: [Leon Li](https://huggingface.co/Leon-Leee)
|
26 |
+
- License: [DeepSeek](https://github.com/deepseek-ai/DeepSeek-Coder/blob/main/LICENSE-MODEL)
|
27 |
+
- Fine-tuned from [deepseek-ai/deepseek-coder-6.7b-base](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-base) with full parameters
|
28 |
+
|
29 |
+
### Training data
|
30 |
+
A mixture of both
|
31 |
+
- samples from several high-quality open-source datasets (read *Acknowledgements*),
|
32 |
+
- our private datasets (already decontaminated with benchmarks).
|
33 |
+
|
34 |
+
### Evaluation
|
35 |
+
To check out our evaluation results: [EvalPlus](https://evalplus.github.io/leaderboard.html)
|
36 |
+
|
37 |
+
### Requirements
|
38 |
+
It should work with the same requirements as DeepSeek-Coder-6.7B
|
39 |
+
```torch>=2.0
|
40 |
+
tokenizers>=0.14.0
|
41 |
+
transformers>=4.35.0
|
42 |
+
accelerate
|
43 |
+
sympy>=1.12
|
44 |
+
pebble
|
45 |
+
timeout-decorator
|
46 |
+
attrdict
|
47 |
+
```
|
48 |
+
|
49 |
+
|
50 |
+
### QuickStart
|
51 |
+
TBD
|
52 |
+
```
|
53 |
+
```
|
54 |
+
|
55 |
+
### Limits
|
56 |
+
|
57 |
+
|
58 |
+
### Acknowledgements
|
59 |
+
- [WizardCoder](https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder): WizardLM-Evol-Instruct V2 datasets
|
60 |
+
- We used a back-up([Leon-Leee/wizardlm_evol_instruct_v2_196K_backuped](https://huggingface.co/datasets/Leon-Leee/wizardlm_evol_instruct_v2_196K_backuped)) since this dataset has been deleted.
|
61 |
+
- [Magicoder](https://github.com/ise-uiuc/magicoder/): [Magicoder-Evol-Instruct-110K](https://huggingface.co/datasets/ise-uiuc/Magicoder-Evol-Instruct-110K) from theblackcat102/evol-codealpaca-v1(https://huggingface.co/datasets/theblackcat102/evol-codealpaca-v1)
|
62 |
+
- [Eurus](https://github.com/OpenBMB/Eurus): reasoning enhancement dataset of [openbmb/UltraInteract_sft](https://huggingface.co/datasets/openbmb/UltraInteract_sft)
|
63 |
+
- [OpenCoderInterpreter](https://opencodeinterpreter.github.io/): [m-a-p/Code-Feedback](https://huggingface.co/datasets/m-a-p/Code-Feedback)
|