Spaces:
Running
Running
Commit
·
63ee5d3
1
Parent(s):
6d3a2f2
No FP8
Browse files
app.py
CHANGED
@@ -42,7 +42,7 @@ def load_model():
|
|
42 |
# tokenizer = AutoTokenizer.from_pretrained(model_id, token=True)
|
43 |
# model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
|
44 |
# model_id = "deepseek-ai/deepseek-llm-7b-chat"
|
45 |
-
model_id = "deepseek-ai/DeepSeek-
|
46 |
|
47 |
quantization_config = FineGrainedFP8Config()
|
48 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
@@ -52,8 +52,8 @@ def load_model():
|
|
52 |
# torch_dtype=torch.float32
|
53 |
device_map="auto",
|
54 |
torch_dtype=torch.float16,
|
55 |
-
|
56 |
-
|
57 |
)
|
58 |
# model.to("cpu")
|
59 |
return tokenizer, model
|
|
|
42 |
# tokenizer = AutoTokenizer.from_pretrained(model_id, token=True)
|
43 |
# model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
|
44 |
# model_id = "deepseek-ai/deepseek-llm-7b-chat"
|
45 |
+
model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
|
46 |
|
47 |
quantization_config = FineGrainedFP8Config()
|
48 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
|
|
52 |
# torch_dtype=torch.float32
|
53 |
device_map="auto",
|
54 |
torch_dtype=torch.float16,
|
55 |
+
# quantization_config=quantization_config,
|
56 |
+
trust_remote_code = True
|
57 |
)
|
58 |
# model.to("cpu")
|
59 |
return tokenizer, model
|