anisrashidov commited on
Commit
712ea2e
·
verified ·
1 Parent(s): fa60f2d

Update crawler.py

Browse files
Files changed (1) hide show
  1. crawler.py +10 -10
crawler.py CHANGED
@@ -93,17 +93,17 @@ def get_answers(results_a_elements, query):
93
  print("Result links extracted: ", len(results_a_elements))
94
 
95
  # Limit the number of parallel processes for better resource management
96
- max_processes = max(1, int(multiprocessing.cpu_count() * 0.5))
97
 
98
- with multiprocessing.Pool(processes=max_processes) as pool:
99
- results = pool.map(process_url, results_a_elements)
100
-
101
- # results = []
102
- # # answer_count = 0
103
- # for url in tqdm.tqdm(results_a_elements):
104
- # res = process_url(url)
105
- # results.append(res)
106
- # answer_count += len(res['answers'])
107
  return results
108
 
109
  def get_search_results(query, num_pages):
 
93
  print("Result links extracted: ", len(results_a_elements))
94
 
95
  # Limit the number of parallel processes for better resource management
96
+ # max_processes = max(1, int(multiprocessing.cpu_count() * 0.5))
97
 
98
+ # with multiprocessing.Pool(processes=max_processes) as pool:
99
+ # results = pool.map(process_url, results_a_elements)
100
+
101
+ results = []
102
+ # answer_count = 0
103
+ for url in tqdm.tqdm(results_a_elements):
104
+ res = process_url(url)
105
+ results.append(res)
106
+ answer_count += len(res['answers'])
107
  return results
108
 
109
  def get_search_results(query, num_pages):