Spaces:
Sleeping
Sleeping
import concurrent.futures | |
import requests | |
import tldextract | |
from pathlib import Path | |
from utils.enver import enver | |
from utils.logger import logger | |
from networks.filepath_converter import UrlToFilepathConverter | |
from networks.network_configs import IGNORE_HOSTS, REQUESTS_HEADERS | |
class WebpageFetcher: | |
def __init__(self): | |
self.enver = enver | |
self.enver.set_envs(proxies=True) | |
self.filepath_converter = UrlToFilepathConverter() | |
def is_ignored_host(self, url): | |
self.host = tldextract.extract(url).registered_domain | |
if self.host in IGNORE_HOSTS: | |
return True | |
else: | |
return False | |
def send_request(self): | |
try: | |
self.request_response = requests.get( | |
url=self.url, | |
headers=REQUESTS_HEADERS, | |
proxies=self.enver.requests_proxies, | |
timeout=15, | |
) | |
except: | |
logger.warn(f"Failed to fetch: [{self.url}]") | |
self.request_response = None | |
def save_response(self): | |
if not self.html_path.exists(): | |
self.html_path.parent.mkdir(parents=True, exist_ok=True) | |
logger.success(f"Saving to: [{self.html_path}]") | |
if self.request_response is None: | |
return | |
else: | |
with open(self.html_path, "wb") as wf: | |
wf.write(self.request_response.content) | |
def fetch(self, url, overwrite=False, output_parent=None): | |
self.url = url | |
logger.note(f"Fetching: [{self.url}]") | |
self.html_path = self.filepath_converter.convert(self.url, parent=output_parent) | |
if self.is_ignored_host(self.url): | |
logger.warn(f"Ignore host: [{self.host}]") | |
return self.html_path | |
if self.html_path.exists() and not overwrite: | |
logger.success(f"HTML existed: [{self.html_path}]") | |
else: | |
self.send_request() | |
self.save_response() | |
return self.html_path | |
class BatchWebpageFetcher: | |
def __init__(self): | |
self.done_count = 0 | |
self.total_count = 0 | |
self.url_and_html_path_list = [] | |
def fecth_single_webpage(self, url, overwrite=False, output_parent=None): | |
webpage_fetcher = WebpageFetcher() | |
html_path = webpage_fetcher.fetch( | |
url=url, overwrite=overwrite, output_parent=output_parent | |
) | |
self.url_and_html_path_list.append({"url": url, "html_path": html_path}) | |
self.done_count += 1 | |
logger.success(f"> [{self.done_count}/{self.total_count}] Fetched: {url}") | |
def fetch(self, urls, overwrite=False, output_parent=None): | |
self.urls = urls | |
self.total_count = len(self.urls) | |
with concurrent.futures.ThreadPoolExecutor() as executor: | |
futures = [ | |
executor.submit( | |
self.fecth_single_webpage, | |
url=url, | |
overwrite=overwrite, | |
output_parent=output_parent, | |
) | |
for url in urls | |
] | |
for idx, future in enumerate(concurrent.futures.as_completed(futures)): | |
result = future.result() | |
return self.url_and_html_path_list | |
if __name__ == "__main__": | |
urls = [ | |
"https://stackoverflow.com/questions/295135/turn-a-string-into-a-valid-filename", | |
"https://www.liaoxuefeng.com/wiki/1016959663602400/1017495723838528", | |
"https://docs.python.org/zh-cn/3/tutorial/interpreter.html", | |
] | |
batch_webpage_fetcher = BatchWebpageFetcher() | |
batch_webpage_fetcher.fetch( | |
urls=urls, overwrite=True, output_parent="python tutorials" | |
) | |