|
import gradio as gr |
|
import requests |
|
import json |
|
from datetime import datetime, timedelta |
|
import pandas as pd |
|
|
|
API_KEY = "V38CNn4HXpLtynJQyOeoUensTEYoFy8PBUxKpDqAW1pawT1vfJ2BWtPQ98h6" |
|
|
|
MAJOR_COUNTRIES = [ |
|
"United States", "United Kingdom", "Canada", "Australia", "Germany", |
|
"France", "Japan", "South Korea", "China", "India", |
|
"Brazil", "Mexico", "Russia", "Italy", "Spain", |
|
"Netherlands", "Sweden", "Switzerland", "Norway", "Denmark", |
|
"Finland", "Belgium", "Austria", "New Zealand", "Ireland", |
|
"Singapore", "Hong Kong", "Israel", "United Arab Emirates", "Saudi Arabia", |
|
"South Africa", "Turkey", "Egypt", "Poland", "Czech Republic", |
|
"Hungary", "Greece", "Portugal", "Argentina", "Chile", |
|
"Colombia", "Peru", "Venezuela", "Thailand", "Malaysia", |
|
"Indonesia", "Philippines", "Vietnam", "Pakistan", "Bangladesh" |
|
] |
|
|
|
def search_serphouse(query, country, page, num_result): |
|
url = "https://api.serphouse.com/serp/live" |
|
|
|
now = datetime.utcnow() |
|
yesterday = now - timedelta(days=1) |
|
date_range = f"{yesterday.strftime('%Y-%m-%d')},{now.strftime('%Y-%m-%d')}" |
|
|
|
payload = { |
|
"data": { |
|
"q": query, |
|
"domain": "google.com", |
|
"loc": country, |
|
"lang": "en", |
|
"device": "desktop", |
|
"serp_type": "news", |
|
"page": str(page), |
|
"verbatim": "1", |
|
"num": str(num_result), |
|
"date_range": date_range |
|
} |
|
} |
|
|
|
headers = { |
|
"accept": "application/json", |
|
"content-type": "application/json", |
|
"authorization": f"Bearer {API_KEY}" |
|
} |
|
|
|
try: |
|
response = requests.post(url, json=payload, headers=headers) |
|
response.raise_for_status() |
|
return response.json() |
|
except requests.RequestException as e: |
|
error_msg = f"Error: {str(e)}" |
|
if response.text: |
|
error_msg += f"\nResponse content: {response.text}" |
|
return {"error": error_msg} |
|
|
|
def format_results(results): |
|
debug_info = f"Raw API Response:\n{json.dumps(results, indent=2, ensure_ascii=False)}\n\n" |
|
|
|
try: |
|
if isinstance(results, dict) and "error" in results: |
|
return pd.DataFrame({"Error": [results["error"]]}), debug_info |
|
|
|
if not isinstance(results, dict): |
|
raise ValueError("๊ฒฐ๊ณผ๊ฐ ์ฌ์ ํ์์ด ์๋๋๋ค.") |
|
|
|
organic_results = results.get("organic_results", []) |
|
debug_info += f"์ ๊ธฐ์ ๊ฒฐ๊ณผ ์: {len(organic_results)}\n" |
|
|
|
if not organic_results: |
|
return pd.DataFrame({"Message": ["๊ฒ์ ๊ฒฐ๊ณผ๊ฐ ์์ต๋๋ค."]}), debug_info |
|
|
|
formatted_results = [] |
|
for result in organic_results: |
|
formatted_results.append({ |
|
"์ ๋ชฉ": result.get("title", "์ ๋ชฉ ์์"), |
|
"๋งํฌ": result.get("link", "#"), |
|
"๋ด์ฉ": result.get("snippet", "๋ด์ฉ ์์"), |
|
"์ถ์ฒ": result.get("source", "์ ์ ์์"), |
|
"์๊ฐ": result.get("date", "์ ์ ์๋ ์๊ฐ") |
|
}) |
|
|
|
return pd.DataFrame(formatted_results), debug_info |
|
|
|
except Exception as e: |
|
error_message = f"๊ฒฐ๊ณผ ์ฒ๋ฆฌ ์ค ์ค๋ฅ ๋ฐ์: {str(e)}" |
|
debug_info += f"Error: {error_message}\n" |
|
debug_info += f"Results structure: {type(results)}\n" |
|
if isinstance(results, dict): |
|
debug_info += f"Results keys: {results.keys()}\n" |
|
return pd.DataFrame({"Error": [error_message]}), debug_info |
|
|
|
def serphouse_search(query, country, page, num_result): |
|
results = search_serphouse(query, country, page, num_result) |
|
df_results, debug_info = format_results(results) |
|
return df_results, debug_info |
|
|
|
css = """ |
|
footer { |
|
visibility: hidden; |
|
} |
|
""" |
|
|
|
iface = gr.Interface( |
|
fn=serphouse_search, |
|
inputs=[ |
|
gr.Textbox(label="๊ฒ์์ด"), |
|
gr.Dropdown(MAJOR_COUNTRIES, label="๊ตญ๊ฐ"), |
|
gr.Slider(1, 10, 1, label="ํ์ด์ง"), |
|
gr.Slider(1, 100, 10, label="๊ฒฐ๊ณผ ์") |
|
], |
|
outputs=[ |
|
gr.Dataframe(label="๋ด์ค ๊ฒฐ๊ณผ"), |
|
gr.Textbox(label="๋๋ฒ๊ทธ ์ ๋ณด", lines=10) |
|
], |
|
title="24์๊ฐ ์ด๋ด ๋ด์ค ๊ฒ์ ์ธํฐํ์ด์ค", |
|
description="๊ฒ์์ด๋ฅผ ์
๋ ฅํ๊ณ ๊ตญ๊ฐ๋ฅผ ์ ํํ์ฌ 24์๊ฐ ์ด๋ด์ ๋ด์ค ๊ฒฐ๊ณผ๋ฅผ ๊ฐ์ ธ์ต๋๋ค.", |
|
theme="Nymbo/Nymbo_Theme", |
|
css=css |
|
) |
|
|
|
|
|
iface.launch(auth=("gini","pick")) |