Spaces:
Running
Running
Commit
·
c8669d6
1
Parent(s):
2427a02
Refactor API key handling and update dependencies; remove unused PDF parsing tool
Browse files- .gitignore +2 -1
- main.py +6 -3
- requirements.txt +5 -1
- tools.py +0 -19
.gitignore
CHANGED
@@ -1 +1,2 @@
|
|
1 |
-
__pycache__
|
|
|
|
1 |
+
__pycache__
|
2 |
+
.env
|
main.py
CHANGED
@@ -1,8 +1,11 @@
|
|
1 |
from math_summarizer import generate_math_summary
|
2 |
from nlp_summarizer import generate_nlp_summary_and_mindmap
|
3 |
-
from tools import parse_pdf
|
4 |
import openai
|
5 |
-
import
|
|
|
|
|
|
|
|
|
6 |
|
7 |
def create_client(api_key):
|
8 |
client = openai.OpenAI(
|
@@ -30,6 +33,6 @@ def generate_summary(client, corpus):
|
|
30 |
return response
|
31 |
|
32 |
def main(corpus):
|
33 |
-
client = create_client(
|
34 |
response = generate_summary(client, corpus)
|
35 |
return response
|
|
|
1 |
from math_summarizer import generate_math_summary
|
2 |
from nlp_summarizer import generate_nlp_summary_and_mindmap
|
|
|
3 |
import openai
|
4 |
+
import dotenv
|
5 |
+
import os
|
6 |
+
|
7 |
+
dotenv.load_dotenv()
|
8 |
+
API_KEY = os.getenv('API_KEY')
|
9 |
|
10 |
def create_client(api_key):
|
11 |
client = openai.OpenAI(
|
|
|
33 |
return response
|
34 |
|
35 |
def main(corpus):
|
36 |
+
client = create_client(API_KEY)
|
37 |
response = generate_summary(client, corpus)
|
38 |
return response
|
requirements.txt
CHANGED
@@ -1 +1,5 @@
|
|
1 |
-
openai
|
|
|
|
|
|
|
|
|
|
1 |
+
openai==1.57.3
|
2 |
+
gradio==5.8.0
|
3 |
+
python-dotenv==1.0.1
|
4 |
+
sumy==0.11.0
|
5 |
+
nltk==3.9.1
|
tools.py
DELETED
@@ -1,19 +0,0 @@
|
|
1 |
-
from langchain_community.document_loaders import PyPDFLoader
|
2 |
-
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
3 |
-
|
4 |
-
def parse_pdf(pdf_path):
|
5 |
-
loader = PyPDFLoader(pdf_path)
|
6 |
-
pages = loader.load_and_split()
|
7 |
-
all_text = " ".join([page.page_content for page in pages])
|
8 |
-
start_index = all_text.find("ABSTRACT")
|
9 |
-
end_index = all_text.find("REFERENCES")
|
10 |
-
if start_index != -1 and end_index != -1 and start_index < end_index:
|
11 |
-
relevant_text = all_text[start_index:end_index]
|
12 |
-
else:
|
13 |
-
relevant_text = all_text
|
14 |
-
text_splitter = RecursiveCharacterTextSplitter(chunk_size=250, chunk_overlap=50)
|
15 |
-
text_list = text_splitter.split_text(relevant_text)
|
16 |
-
research_paper_text = "".join(text_list)
|
17 |
-
length_of_research_paper = len(research_paper_text)
|
18 |
-
print(f"Length of Research Paper: {length_of_research_paper}")
|
19 |
-
return research_paper_text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|