File size: 5,000 Bytes
09ebcd0 eec4d01 efb5f18 09ebcd0 6097fb2 1b298f9 34db24b 6097fb2 1b298f9 8415ea5 6097fb2 2efa0ec eec4d01 8415ea5 eec4d01 1b298f9 6097fb2 eec4d01 6097fb2 ba53ac8 6097fb2 ba53ac8 1b298f9 a31c3ed 8415ea5 3adcf52 8415ea5 3adcf52 8415ea5 3adcf52 73ed8f5 3adcf52 8415ea5 3adcf52 482bc99 6097fb2 d0b4b99 ae814df a4f4206 73ed8f5 ba53ac8 dc63a3a c5cb552 dc63a3a c5cb552 dc63a3a c5cb552 dc63a3a ba53ac8 dc63a3a efb5f18 ae814df 158b524 8415ea5 ae814df 73ed8f5 68a8b0c 43da967 716e5e1 ba53ac8 6097fb2 ab30506 6097fb2 43da967 ad0f104 d0b4b99 73ed8f5 ae814df d0b4b99 43da967 d0b4b99 43da967 ad0f104 951a64a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 |
import gradio as gr
import requests
import json
from datetime import datetime, timedelta
API_KEY = "V38CNn4HXpLtynJQyOeoUensTEYoFy8PBUxKpDqAW1pawT1vfJ2BWtPQ98h6"
MAJOR_COUNTRIES = [
"United States", "United Kingdom", "Canada", "Australia", "Germany",
"France", "Japan", "South Korea", "China", "India",
"Brazil", "Mexico", "Russia", "Italy", "Spain",
"Netherlands", "Sweden", "Switzerland", "Norway", "Denmark",
"Finland", "Belgium", "Austria", "New Zealand", "Ireland",
"Singapore", "Hong Kong", "Israel", "United Arab Emirates", "Saudi Arabia",
"South Africa", "Turkey", "Egypt", "Poland", "Czech Republic",
"Hungary", "Greece", "Portugal", "Argentina", "Chile",
"Colombia", "Peru", "Venezuela", "Thailand", "Malaysia",
"Indonesia", "Philippines", "Vietnam", "Pakistan", "Bangladesh"
]
def search_serphouse(query, country, page, num_result):
url = "https://api.serphouse.com/serp/live"
payload = {
"data": {
"q": query,
"domain": "google.com",
"loc": country,
"lang": "en",
"device": "desktop",
"serp_type": "news",
"page": str(page),
"verbatim": "1",
"num": str(num_result)
}
}
headers = {
"accept": "application/json",
"content-type": "application/json",
"authorization": f"Bearer {API_KEY}"
}
try:
response = requests.post(url, json=payload, headers=headers)
response.raise_for_status()
return response.json()
except requests.RequestException as e:
return f"Error: {str(e)}"
def is_recent_news(time_str):
if not time_str:
return False
time_parts = time_str.lower().split()
if len(time_parts) < 2:
return False
try:
value = int(time_parts[0])
unit = time_parts[1]
if unit in ['minute', 'minutes', 'hour', 'hours']:
return True
elif unit in ['day', 'days']:
return value <= 1
except ValueError:
return False
return False
def format_results(results):
all_results = "## All News Results\n\n"
recent_results = "## Recent News Results (Within 1 Day)\n\n"
debug_info = "## Debug Information\n\n"
debug_info += f"Raw API Response:\n```json\n{json.dumps(results, indent=2)}\n```\n\n"
try:
if not isinstance(results, dict):
raise ValueError("Results is not a dictionary")
if "results" not in results:
raise ValueError("No 'results' key in the response")
if "news" not in results["results"]:
raise ValueError("No 'news' key in results")
news_results = results["results"]["news"]
debug_info += f"Number of news results: {len(news_results)}\n\n"
for result in news_results:
title = result.get("title", "No Title")
url = result.get("url", "#")
snippet = result.get("snippet", "No Snippet")
channel = result.get("channel", "Unknown")
time_str = result.get("time", "Unknown time")
is_recent = is_recent_news(time_str)
debug_info += f"Article: {title}\nTime: {time_str}, Is Recent: {is_recent}\n\n"
article_info = f"""
### [{title}]({url})
{snippet}
**Source:** {channel} - {time_str}
---
"""
all_results += article_info
if is_recent:
recent_results += article_info
if recent_results == "## Recent News Results (Within 1 Day)\n\n":
recent_results += "*No recent news results found within 1 day.*\n\n"
except Exception as e:
error_message = f"Error processing results: {str(e)}"
debug_info += error_message + "\n"
all_results = error_message + "\n\n"
recent_results = error_message + "\n\n"
return all_results, recent_results, debug_info
def serphouse_search(query, country, page, num_result):
results = search_serphouse(query, country, page, num_result)
all_results, recent_results, debug_info = format_results(results)
return all_results, recent_results, debug_info
css = """
footer {
visibility: hidden;
}
"""
iface = gr.Interface(
fn=serphouse_search,
inputs=[
gr.Textbox(label="Search Query"),
gr.Dropdown(MAJOR_COUNTRIES, label="Country"),
gr.Slider(1, 10, 1, label="Page"),
gr.Slider(1, 100, 10, label="Number of Results")
],
outputs=[
gr.Markdown(label="All Results"),
gr.Markdown(label="Recent Results (Within 1 Day)"),
gr.Markdown(label="Debug Information")
],
title="SERPHouse News Search Interface",
description="Enter your search query and select a country to get news results from the SERPHouse API. Recent results (within 1 day) are shown separately.",
theme="Nymbo/Nymbo_Theme",
css=css
)
iface.launch()
|