Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -6,6 +6,10 @@ from huggingface_hub import HfFolder, login
|
|
6 |
from io import StringIO
|
7 |
import os
|
8 |
from flask import Flask, request, jsonify
|
|
|
|
|
|
|
|
|
9 |
|
10 |
# Load GPT-2 model and tokenizer
|
11 |
|
@@ -65,6 +69,7 @@ def format_prompt(description, columns):
|
|
65 |
|
66 |
API_URL = "https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3.1-8B"
|
67 |
|
|
|
68 |
generation_params = {
|
69 |
"top_p": 0.90,
|
70 |
"temperature": 0.8,
|
@@ -86,6 +91,7 @@ def generate_synthetic_data(description, columns):
|
|
86 |
else:
|
87 |
raise ValueError("Unexpected response format or missing 'generated_text' key")
|
88 |
else:
|
|
|
89 |
raise ValueError(f"API request failed with status code {response.status_code}: {response.text}")
|
90 |
|
91 |
def generate_large_synthetic_data(description, columns, num_rows=1000, rows_per_generation=100):
|
|
|
6 |
from io import StringIO
|
7 |
import os
|
8 |
from flask import Flask, request, jsonify
|
9 |
+
from huggingface_hub import HfFolder
|
10 |
+
|
11 |
+
print(HfFolder.get_token())
|
12 |
+
login()
|
13 |
|
14 |
# Load GPT-2 model and tokenizer
|
15 |
|
|
|
69 |
|
70 |
API_URL = "https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3.1-8B"
|
71 |
|
72 |
+
|
73 |
generation_params = {
|
74 |
"top_p": 0.90,
|
75 |
"temperature": 0.8,
|
|
|
91 |
else:
|
92 |
raise ValueError("Unexpected response format or missing 'generated_text' key")
|
93 |
else:
|
94 |
+
print(f"Error details: {response.text}")
|
95 |
raise ValueError(f"API request failed with status code {response.status_code}: {response.text}")
|
96 |
|
97 |
def generate_large_synthetic_data(description, columns, num_rows=1000, rows_per_generation=100):
|