import os import re import sys from urllib import request as urlrequest from constants import CACHE_DIR_MULTIHMR, SMPLX_DIR, MEAN_PARAMS def _progress_bar(count, total): """Report download progress. Credit: https://stackoverflow.com/questions/3173320/text-progress-bar-in-the-console/27871113 """ bar_len = 60 filled_len = int(round(bar_len * count / float(total))) percents = round(100.0 * count / float(total), 1) bar = "=" * filled_len + "-" * (bar_len - filled_len) sys.stdout.write( " [{}] {}% of {:.1f}MB file \r".format(bar, percents, total / 1024 / 1024) ) sys.stdout.flush() if count >= total: sys.stdout.write("\n") def download_url(url, dst_file_path, chunk_size=8192, progress_hook=_progress_bar): """Download url and write it to dst_file_path. Credit: https://stackoverflow.com/questions/2028517/python-urllib2-progress-hook """ # url = url + "?dl=1" if "dropbox" in url else url req = urlrequest.Request(url) response = urlrequest.urlopen(req) total_size = response.info().get("Content-Length") if total_size is None: raise ValueError("Cannot determine size of download from {}".format(url)) total_size = int(total_size.strip()) bytes_so_far = 0 with open(dst_file_path, "wb") as f: while 1: chunk = response.read(chunk_size) bytes_so_far += len(chunk) if not chunk: break if progress_hook: progress_hook(bytes_so_far, total_size) f.write(chunk) return bytes_so_far def cache_url(url_or_file, cache_file_path, download=True): """Download the file specified by the URL to the cache_dir and return the path to the cached file. If the argument is not a URL, simply return it as is. """ is_url = re.match(r"^(?:http)s?://", url_or_file, re.IGNORECASE) is not None if not is_url: return url_or_file url = url_or_file if os.path.exists(cache_file_path): return cache_file_path cache_file_dir = os.path.dirname(cache_file_path) if not os.path.exists(cache_file_dir): os.makedirs(cache_file_dir) if download: print("Downloading remote file {} to {}".format(url, cache_file_path)) download_url(url, cache_file_path) return cache_file_path def download_models(folder=CACHE_DIR_MULTIHMR): """Download checkpoints and files for running inference. """ import os os.makedirs(folder, exist_ok=True) download_files = { "hmr2_data.tar.gz" : ["https://people.eecs.berkeley.edu/~jathushan/projects/4dhumans/hmr2_data.tar.gz", folder], } for file_name, url in download_files.items(): output_path = os.path.join(url[1], file_name) if not os.path.exists(output_path): print("Downloading file: " + file_name) # output = gdown.cached_download(url[0], output_path, fuzzy=True) output = cache_url(url[0], output_path) assert os.path.exists(output_path), f"{output} does not exist" # if ends with tar.gz, tar -xzf if file_name.endswith(".tar.gz"): print("Extracting file: " + file_name) os.system("tar -xvf " + output_path + " -C " + url[1]) def check_smplx_exists(): import os candidates = [ f'{SMPLX_DIR}/data/smplx/SMPLX_NEUTRAL.npz', f'{MEAN_PARAMS}' ] candidates_exist = [os.path.exists(c) for c in candidates] if not any(candidates_exist): raise FileNotFoundError(f"SMPLX model not found. Please download it from https://smplify.is.tue.mpg.de/ and place it at {candidates[1]}") return True