Spaces:
Running
Running
Commit
·
3fa478b
1
Parent(s):
697b1f4
Works?
Browse files- app.py +2 -2
- requirements.txt +1 -1
app.py
CHANGED
@@ -44,7 +44,7 @@ def load_model():
|
|
44 |
# model_id = "deepseek-ai/deepseek-llm-7b-chat"
|
45 |
# model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
|
46 |
|
47 |
-
model_id = "deepseek-ai/DeepSeek-
|
48 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
49 |
model = AutoModelForCausalLM.from_pretrained(
|
50 |
model_id,
|
@@ -53,7 +53,7 @@ def load_model():
|
|
53 |
device_map="auto",
|
54 |
torch_dtype=torch.float16,
|
55 |
# quantization_config=quantization_config,
|
56 |
-
# attn_implementation="flash_attention_2",
|
57 |
trust_remote_code = True
|
58 |
)
|
59 |
# model.to("cpu")
|
|
|
44 |
# model_id = "deepseek-ai/deepseek-llm-7b-chat"
|
45 |
# model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
|
46 |
|
47 |
+
model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
|
48 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
49 |
model = AutoModelForCausalLM.from_pretrained(
|
50 |
model_id,
|
|
|
53 |
device_map="auto",
|
54 |
torch_dtype=torch.float16,
|
55 |
# quantization_config=quantization_config,
|
56 |
+
# attn_implementation="flash_attention_2",
|
57 |
trust_remote_code = True
|
58 |
)
|
59 |
# model.to("cpu")
|
requirements.txt
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
streamlit
|
2 |
-
|
3 |
torch
|
4 |
accelerate
|
|
|
1 |
streamlit
|
2 |
+
transformers
|
3 |
torch
|
4 |
accelerate
|