Shreyas94 commited on
Commit
5ccfc6a
·
verified ·
1 Parent(s): e28d7cf

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -10
app.py CHANGED
@@ -2,12 +2,15 @@ import gradio as gr
2
  from transformers import pipeline
3
  from bs4 import BeautifulSoup
4
  import requests
5
- from googlesearch import search
6
  import logging
 
7
 
8
  # Configure logging
9
  logging.basicConfig(level=logging.DEBUG)
10
 
 
 
 
11
  def summarize_news(query, num_results=3):
12
  logging.debug(f"Query received: {query}")
13
  logging.debug(f"Number of results requested: {num_results}")
@@ -18,15 +21,18 @@ def summarize_news(query, num_results=3):
18
 
19
  # Search for news articles
20
  logging.debug("Searching for news articles...")
21
- search_results = list(search(query, num_results=num_results))
22
  articles = []
 
 
 
 
 
 
 
 
23
 
24
- logging.debug(f"Search results: {search_results}")
25
-
26
- for url in search_results:
27
- try:
28
- logging.debug(f"Fetching content from URL: {url}")
29
  # Fetch the content of the news article
 
30
  r = requests.get(url)
31
  soup = BeautifulSoup(r.text, 'html.parser')
32
  results = soup.find_all(['h1', 'p'])
@@ -65,9 +71,8 @@ def summarize_news(query, num_results=3):
65
  articles.append((url, summary_text))
66
 
67
  logging.debug(f"Summary for URL {url}: {summary_text}")
68
- except Exception as e:
69
- logging.error(f"Error processing URL {url}: {e}")
70
- continue
71
 
72
  logging.debug(f"Final summarized articles: {articles}")
73
  return format_output(articles)
 
2
  from transformers import pipeline
3
  from bs4 import BeautifulSoup
4
  import requests
 
5
  import logging
6
+ from newsapi import NewsApiClient
7
 
8
  # Configure logging
9
  logging.basicConfig(level=logging.DEBUG)
10
 
11
+ # Initialize the News API client
12
+ newsapi = NewsApiClient(api_key='5ab7bb1aaceb41b8993db03477098aad')
13
+
14
  def summarize_news(query, num_results=3):
15
  logging.debug(f"Query received: {query}")
16
  logging.debug(f"Number of results requested: {num_results}")
 
21
 
22
  # Search for news articles
23
  logging.debug("Searching for news articles...")
 
24
  articles = []
25
+ try:
26
+ news_results = newsapi.get_everything(q=query, language='en', page_size=num_results)
27
+ logging.debug(f"Search results: {news_results}")
28
+
29
+ for article in news_results['articles']:
30
+ url = article['url']
31
+ title = article['title']
32
+ description = article['description']
33
 
 
 
 
 
 
34
  # Fetch the content of the news article
35
+ logging.debug(f"Fetching content from URL: {url}")
36
  r = requests.get(url)
37
  soup = BeautifulSoup(r.text, 'html.parser')
38
  results = soup.find_all(['h1', 'p'])
 
71
  articles.append((url, summary_text))
72
 
73
  logging.debug(f"Summary for URL {url}: {summary_text}")
74
+ except Exception as e:
75
+ logging.error(f"Error fetching news articles: {e}")
 
76
 
77
  logging.debug(f"Final summarized articles: {articles}")
78
  return format_output(articles)