Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -6,11 +6,16 @@ from huggingface_hub import HfFolder, login
|
|
6 |
from io import StringIO
|
7 |
|
8 |
# Load GPT-2 model and tokenizer
|
9 |
-
|
|
|
|
|
|
|
|
|
|
|
10 |
model_gpt2 = GPT2LMHeadModel.from_pretrained('gpt2')
|
11 |
|
12 |
# Create a pipeline for text generation using GPT-2
|
13 |
-
text_generator = pipeline("text-generation", model=model_gpt2)
|
14 |
|
15 |
# Load the LLaMA tokenizer
|
16 |
tokenizer_llama = LlamaTokenizer.from_pretrained("meta-llama/Meta-Llama-3.1-8B")
|
|
|
6 |
from io import StringIO
|
7 |
|
8 |
# Load GPT-2 model and tokenizer
|
9 |
+
|
10 |
+
# Set environment variable to avoid floating-point errors
|
11 |
+
os.environ['TF_ENABLE_ONEDNN_OPTS'] = '0'
|
12 |
+
|
13 |
+
# Load the tokenizer and model
|
14 |
+
tokenizer = GPT2Tokenizer.from_pretrained('gpt2')
|
15 |
model_gpt2 = GPT2LMHeadModel.from_pretrained('gpt2')
|
16 |
|
17 |
# Create a pipeline for text generation using GPT-2
|
18 |
+
text_generator = pipeline("text-generation", model=model_gpt2, tokenizer=tokenizer)
|
19 |
|
20 |
# Load the LLaMA tokenizer
|
21 |
tokenizer_llama = LlamaTokenizer.from_pretrained("meta-llama/Meta-Llama-3.1-8B")
|