Update README.md
Browse files
README.md
CHANGED
|
@@ -1,21 +1,14 @@
|
|
| 1 |
---
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
tasks:
|
| 6 |
-
- text-generation
|
| 7 |
---
|
| 8 |
# Model Card for CodeFuse-DeepSeek-33B
|
| 9 |
-
|
| 10 |
-
<img src="https://modelscope.cn/api/v1/models/codefuse-ai/CodeFuse-DeepSeek-33B/repo?Revision=master&FilePath=LOGO.jpg&View=true" width="800"/>
|
| 11 |
-
<p>
|
| 12 |
|
| 13 |
[[中文]](#chinese) [[English]](#english)
|
| 14 |
|
| 15 |
-
|
| 16 |
-
```bash
|
| 17 |
-
git clone https://www.modelscope.cn/codefuse-ai/CodeFuse-DeepSeek-33B.git
|
| 18 |
-
```
|
| 19 |
|
| 20 |
<a id="english"></a>
|
| 21 |
|
|
@@ -27,7 +20,7 @@ CodeFuse-DeepSeek-33B is a 33B Code-LLM finetuned by QLoRA on multiple code-rela
|
|
| 27 |
|
| 28 |
## News and Updates
|
| 29 |
|
| 30 |
-
🔥🔥🔥 2024-01-12 CodeFuse-DeepSeek-33B has been released,
|
| 31 |
|
| 32 |
🔥🔥 2023-11-10 CodeFuse-CodeGeeX2-6B has been released, achieving a pass@1 (greedy decoding) score of 45.12% on HumanEval, which is a 9.22% increase compared to CodeGeeX2 35.9%.
|
| 33 |
|
|
@@ -39,7 +32,7 @@ CodeFuse-DeepSeek-33B is a 33B Code-LLM finetuned by QLoRA on multiple code-rela
|
|
| 39 |
|
| 40 |
🔥🔥🔥 2023-09-26 We are pleased to announce the release of the [4-bit quantized version](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B-4bits/summary) of [CodeFuse-CodeLlama-34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary). Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric.
|
| 41 |
|
| 42 |
-
🔥🔥🔥 2023-09-11 [CodeFuse-CodeLlama34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary) has
|
| 43 |
|
| 44 |
<br>
|
| 45 |
|
|
@@ -133,10 +126,10 @@ In this format, the system section is optional and the conversation can be eithe
|
|
| 133 |
|
| 134 |
```python
|
| 135 |
import torch
|
| 136 |
-
from
|
| 137 |
|
| 138 |
def load_model_tokenizer(model_path):
|
| 139 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
| 140 |
tokenizer.eos_token = "<|end▁of▁sentence|>"
|
| 141 |
tokenizer.pad_token = "<|end▁of▁sentence|>"
|
| 142 |
tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
|
|
@@ -295,10 +288,10 @@ User prompt...
|
|
| 295 |
|
| 296 |
```python
|
| 297 |
import torch
|
| 298 |
-
from
|
| 299 |
|
| 300 |
def load_model_tokenizer(model_path):
|
| 301 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
| 302 |
tokenizer.eos_token = "<|end▁of▁sentence|>"
|
| 303 |
tokenizer.pad_token = "<|end▁of▁sentence|>"
|
| 304 |
tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
|
|
|
|
| 1 |
---
|
| 2 |
+
license: other
|
| 3 |
+
tasks:
|
| 4 |
+
- code-generation
|
|
|
|
|
|
|
| 5 |
---
|
| 6 |
# Model Card for CodeFuse-DeepSeek-33B
|
| 7 |
+

|
|
|
|
|
|
|
| 8 |
|
| 9 |
[[中文]](#chinese) [[English]](#english)
|
| 10 |
|
| 11 |
+
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
<a id="english"></a>
|
| 14 |
|
|
|
|
| 20 |
|
| 21 |
## News and Updates
|
| 22 |
|
| 23 |
+
🔥🔥🔥 2024-01-12 CodeFuse-DeepSeek-33B has been released, achieving a pass@1 (greedy decoding) score of 78.65% on HumanEval.
|
| 24 |
|
| 25 |
🔥🔥 2023-11-10 CodeFuse-CodeGeeX2-6B has been released, achieving a pass@1 (greedy decoding) score of 45.12% on HumanEval, which is a 9.22% increase compared to CodeGeeX2 35.9%.
|
| 26 |
|
|
|
|
| 32 |
|
| 33 |
🔥🔥🔥 2023-09-26 We are pleased to announce the release of the [4-bit quantized version](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B-4bits/summary) of [CodeFuse-CodeLlama-34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary). Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric.
|
| 34 |
|
| 35 |
+
🔥🔥🔥 2023-09-11 [CodeFuse-CodeLlama34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary) has achieved 74.4% of pass@1 (greedy decoding) on HumanEval, which is SOTA results for openspurced LLMs at present.
|
| 36 |
|
| 37 |
<br>
|
| 38 |
|
|
|
|
| 126 |
|
| 127 |
```python
|
| 128 |
import torch
|
| 129 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
|
| 130 |
|
| 131 |
def load_model_tokenizer(model_path):
|
| 132 |
+
tokenizer = AutoTokenizer.from_pretrained("codefuse-ai/CodeFuse-DeepSeek-33B", trust_remote_code=True, use_fast=False, legacy=False)
|
| 133 |
tokenizer.eos_token = "<|end▁of▁sentence|>"
|
| 134 |
tokenizer.pad_token = "<|end▁of▁sentence|>"
|
| 135 |
tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
|
|
|
|
| 288 |
|
| 289 |
```python
|
| 290 |
import torch
|
| 291 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
|
| 292 |
|
| 293 |
def load_model_tokenizer(model_path):
|
| 294 |
+
tokenizer = AutoTokenizer.from_pretrained("codefuse-ai/CodeFuse-DeepSeek-33B", trust_remote_code=True, use_fast=False, legacy=False)
|
| 295 |
tokenizer.eos_token = "<|end▁of▁sentence|>"
|
| 296 |
tokenizer.pad_token = "<|end▁of▁sentence|>"
|
| 297 |
tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
|