raannakasturi commited on
Commit
c8669d6
·
1 Parent(s): 2427a02

Refactor API key handling and update dependencies; remove unused PDF parsing tool

Browse files
Files changed (4) hide show
  1. .gitignore +2 -1
  2. main.py +6 -3
  3. requirements.txt +5 -1
  4. tools.py +0 -19
.gitignore CHANGED
@@ -1 +1,2 @@
1
- __pycache__
 
 
1
+ __pycache__
2
+ .env
main.py CHANGED
@@ -1,8 +1,11 @@
1
  from math_summarizer import generate_math_summary
2
  from nlp_summarizer import generate_nlp_summary_and_mindmap
3
- from tools import parse_pdf
4
  import openai
5
- import time
 
 
 
 
6
 
7
  def create_client(api_key):
8
  client = openai.OpenAI(
@@ -30,6 +33,6 @@ def generate_summary(client, corpus):
30
  return response
31
 
32
  def main(corpus):
33
- client = create_client("glhf_0d3c695626fec941eeb7914dd0e36da5")
34
  response = generate_summary(client, corpus)
35
  return response
 
1
  from math_summarizer import generate_math_summary
2
  from nlp_summarizer import generate_nlp_summary_and_mindmap
 
3
  import openai
4
+ import dotenv
5
+ import os
6
+
7
+ dotenv.load_dotenv()
8
+ API_KEY = os.getenv('API_KEY')
9
 
10
  def create_client(api_key):
11
  client = openai.OpenAI(
 
33
  return response
34
 
35
  def main(corpus):
36
+ client = create_client(API_KEY)
37
  response = generate_summary(client, corpus)
38
  return response
requirements.txt CHANGED
@@ -1 +1,5 @@
1
- openai
 
 
 
 
 
1
+ openai==1.57.3
2
+ gradio==5.8.0
3
+ python-dotenv==1.0.1
4
+ sumy==0.11.0
5
+ nltk==3.9.1
tools.py DELETED
@@ -1,19 +0,0 @@
1
- from langchain_community.document_loaders import PyPDFLoader
2
- from langchain.text_splitter import RecursiveCharacterTextSplitter
3
-
4
- def parse_pdf(pdf_path):
5
- loader = PyPDFLoader(pdf_path)
6
- pages = loader.load_and_split()
7
- all_text = " ".join([page.page_content for page in pages])
8
- start_index = all_text.find("ABSTRACT")
9
- end_index = all_text.find("REFERENCES")
10
- if start_index != -1 and end_index != -1 and start_index < end_index:
11
- relevant_text = all_text[start_index:end_index]
12
- else:
13
- relevant_text = all_text
14
- text_splitter = RecursiveCharacterTextSplitter(chunk_size=250, chunk_overlap=50)
15
- text_list = text_splitter.split_text(relevant_text)
16
- research_paper_text = "".join(text_list)
17
- length_of_research_paper = len(research_paper_text)
18
- print(f"Length of Research Paper: {length_of_research_paper}")
19
- return research_paper_text