Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -77,7 +77,7 @@ LLAMA_MAX_MAX_NEW_TOKENS = 512
|
|
77 |
LLAMA_DEFAULT_MAX_NEW_TOKENS = 512
|
78 |
LLAMA_MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "1024"))
|
79 |
llama_device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
80 |
-
llama_model_id = "HuggingFaceTB/SmolLM2-
|
81 |
llama_tokenizer = AutoTokenizer.from_pretrained(llama_model_id)
|
82 |
llama_model = AutoModelForCausalLM.from_pretrained(
|
83 |
llama_model_id,
|
|
|
77 |
LLAMA_DEFAULT_MAX_NEW_TOKENS = 512
|
78 |
LLAMA_MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "1024"))
|
79 |
llama_device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
80 |
+
llama_model_id = "HuggingFaceTB/SmolLM2-1.7B-Instruct"
|
81 |
llama_tokenizer = AutoTokenizer.from_pretrained(llama_model_id)
|
82 |
llama_model = AutoModelForCausalLM.from_pretrained(
|
83 |
llama_model_id,
|