Spaces:
Running
Running
from __future__ import annotations | |
import importlib | |
import types | |
import warnings | |
__all__ = ["registry", "get_filesystem_class", "default"] | |
# internal, mutable | |
_registry: dict[str, type] = {} | |
# external, immutable | |
registry = types.MappingProxyType(_registry) | |
default = "file" | |
def register_implementation(name, cls, clobber=False, errtxt=None): | |
"""Add implementation class to the registry | |
Parameters | |
---------- | |
name: str | |
Protocol name to associate with the class | |
cls: class or str | |
if a class: fsspec-compliant implementation class (normally inherits from | |
``fsspec.AbstractFileSystem``, gets added straight to the registry. If a | |
str, the full path to an implementation class like package.module.class, | |
which gets added to known_implementations, | |
so the import is deferred until the filesystem is actually used. | |
clobber: bool (optional) | |
Whether to overwrite a protocol with the same name; if False, will raise | |
instead. | |
errtxt: str (optional) | |
If given, then a failure to import the given class will result in this | |
text being given. | |
""" | |
if isinstance(cls, str): | |
if name in known_implementations and clobber is False: | |
if cls != known_implementations[name]["class"]: | |
raise ValueError( | |
f"Name ({name}) already in the known_implementations and clobber " | |
f"is False" | |
) | |
else: | |
known_implementations[name] = { | |
"class": cls, | |
"err": errtxt or f"{cls} import failed for protocol {name}", | |
} | |
else: | |
if name in registry and clobber is False: | |
if _registry[name] is not cls: | |
raise ValueError( | |
f"Name ({name}) already in the registry and clobber is False" | |
) | |
else: | |
_registry[name] = cls | |
# protocols mapped to the class which implements them. This dict can be | |
# updated with register_implementation | |
known_implementations = { | |
"abfs": { | |
"class": "adlfs.AzureBlobFileSystem", | |
"err": "Install adlfs to access Azure Datalake Gen2 and Azure Blob Storage", | |
}, | |
"adl": { | |
"class": "adlfs.AzureDatalakeFileSystem", | |
"err": "Install adlfs to access Azure Datalake Gen1", | |
}, | |
"arrow_hdfs": { | |
"class": "fsspec.implementations.arrow.HadoopFileSystem", | |
"err": "pyarrow and local java libraries required for HDFS", | |
}, | |
"asynclocal": { | |
"class": "morefs.asyn_local.AsyncLocalFileSystem", | |
"err": "Install 'morefs[asynclocalfs]' to use AsyncLocalFileSystem", | |
}, | |
"az": { | |
"class": "adlfs.AzureBlobFileSystem", | |
"err": "Install adlfs to access Azure Datalake Gen2 and Azure Blob Storage", | |
}, | |
"blockcache": {"class": "fsspec.implementations.cached.CachingFileSystem"}, | |
"box": { | |
"class": "boxfs.BoxFileSystem", | |
"err": "Please install boxfs to access BoxFileSystem", | |
}, | |
"cached": {"class": "fsspec.implementations.cached.CachingFileSystem"}, | |
"dask": { | |
"class": "fsspec.implementations.dask.DaskWorkerFileSystem", | |
"err": "Install dask distributed to access worker file system", | |
}, | |
"data": {"class": "fsspec.implementations.data.DataFileSystem"}, | |
"dbfs": { | |
"class": "fsspec.implementations.dbfs.DatabricksFileSystem", | |
"err": "Install the requests package to use the DatabricksFileSystem", | |
}, | |
"dir": {"class": "fsspec.implementations.dirfs.DirFileSystem"}, | |
"dropbox": { | |
"class": "dropboxdrivefs.DropboxDriveFileSystem", | |
"err": ( | |
'DropboxFileSystem requires "dropboxdrivefs","requests" and "' | |
'"dropbox" to be installed' | |
), | |
}, | |
"dvc": { | |
"class": "dvc.api.DVCFileSystem", | |
"err": "Install dvc to access DVCFileSystem", | |
}, | |
"file": {"class": "fsspec.implementations.local.LocalFileSystem"}, | |
"filecache": {"class": "fsspec.implementations.cached.WholeFileCacheFileSystem"}, | |
"ftp": {"class": "fsspec.implementations.ftp.FTPFileSystem"}, | |
"gcs": { | |
"class": "gcsfs.GCSFileSystem", | |
"err": "Please install gcsfs to access Google Storage", | |
}, | |
"gdrive": { | |
"class": "gdrivefs.GoogleDriveFileSystem", | |
"err": "Please install gdrivefs for access to Google Drive", | |
}, | |
"generic": {"class": "fsspec.generic.GenericFileSystem"}, | |
"git": { | |
"class": "fsspec.implementations.git.GitFileSystem", | |
"err": "Install pygit2 to browse local git repos", | |
}, | |
"github": { | |
"class": "fsspec.implementations.github.GithubFileSystem", | |
"err": "Install the requests package to use the github FS", | |
}, | |
"gs": { | |
"class": "gcsfs.GCSFileSystem", | |
"err": "Please install gcsfs to access Google Storage", | |
}, | |
"hdfs": { | |
"class": "fsspec.implementations.arrow.HadoopFileSystem", | |
"err": "pyarrow and local java libraries required for HDFS", | |
}, | |
"hf": { | |
"class": "huggingface_hub.HfFileSystem", | |
"err": "Install huggingface_hub to access HfFileSystem", | |
}, | |
"http": { | |
"class": "fsspec.implementations.http.HTTPFileSystem", | |
"err": 'HTTPFileSystem requires "requests" and "aiohttp" to be installed', | |
}, | |
"https": { | |
"class": "fsspec.implementations.http.HTTPFileSystem", | |
"err": 'HTTPFileSystem requires "requests" and "aiohttp" to be installed', | |
}, | |
"jlab": { | |
"class": "fsspec.implementations.jupyter.JupyterFileSystem", | |
"err": "Jupyter FS requires requests to be installed", | |
}, | |
"jupyter": { | |
"class": "fsspec.implementations.jupyter.JupyterFileSystem", | |
"err": "Jupyter FS requires requests to be installed", | |
}, | |
"lakefs": { | |
"class": "lakefs_spec.LakeFSFileSystem", | |
"err": "Please install lakefs-spec to access LakeFSFileSystem", | |
}, | |
"libarchive": { | |
"class": "fsspec.implementations.libarchive.LibArchiveFileSystem", | |
"err": "LibArchive requires to be installed", | |
}, | |
"local": {"class": "fsspec.implementations.local.LocalFileSystem"}, | |
"memory": {"class": "fsspec.implementations.memory.MemoryFileSystem"}, | |
"oci": { | |
"class": "ocifs.OCIFileSystem", | |
"err": "Install ocifs to access OCI Object Storage", | |
}, | |
"ocilake": { | |
"class": "ocifs.OCIFileSystem", | |
"err": "Install ocifs to access OCI Data Lake", | |
}, | |
"oss": { | |
"class": "ossfs.OSSFileSystem", | |
"err": "Install ossfs to access Alibaba Object Storage System", | |
}, | |
"reference": {"class": "fsspec.implementations.reference.ReferenceFileSystem"}, | |
"root": { | |
"class": "fsspec_xrootd.XRootDFileSystem", | |
"err": ( | |
"Install fsspec-xrootd to access xrootd storage system. " | |
"Note: 'root' is the protocol name for xrootd storage systems, " | |
"not referring to root directories" | |
), | |
}, | |
"s3": {"class": "s3fs.S3FileSystem", "err": "Install s3fs to access S3"}, | |
"s3a": {"class": "s3fs.S3FileSystem", "err": "Install s3fs to access S3"}, | |
"sftp": { | |
"class": "fsspec.implementations.sftp.SFTPFileSystem", | |
"err": 'SFTPFileSystem requires "paramiko" to be installed', | |
}, | |
"simplecache": {"class": "fsspec.implementations.cached.SimpleCacheFileSystem"}, | |
"smb": { | |
"class": "fsspec.implementations.smb.SMBFileSystem", | |
"err": 'SMB requires "smbprotocol" or "smbprotocol[kerberos]" installed', | |
}, | |
"ssh": { | |
"class": "fsspec.implementations.sftp.SFTPFileSystem", | |
"err": 'SFTPFileSystem requires "paramiko" to be installed', | |
}, | |
"tar": {"class": "fsspec.implementations.tar.TarFileSystem"}, | |
"wandb": {"class": "wandbfs.WandbFS", "err": "Install wandbfs to access wandb"}, | |
"webdav": { | |
"class": "webdav4.fsspec.WebdavFileSystem", | |
"err": "Install webdav4 to access WebDAV", | |
}, | |
"webhdfs": { | |
"class": "fsspec.implementations.webhdfs.WebHDFS", | |
"err": 'webHDFS access requires "requests" to be installed', | |
}, | |
"zip": {"class": "fsspec.implementations.zip.ZipFileSystem"}, | |
} | |
assert list(known_implementations) == sorted( | |
known_implementations | |
), "Not in alphabetical order" | |
def get_filesystem_class(protocol): | |
"""Fetch named protocol implementation from the registry | |
The dict ``known_implementations`` maps protocol names to the locations | |
of classes implementing the corresponding file-system. When used for the | |
first time, appropriate imports will happen and the class will be placed in | |
the registry. All subsequent calls will fetch directly from the registry. | |
Some protocol implementations require additional dependencies, and so the | |
import may fail. In this case, the string in the "err" field of the | |
``known_implementations`` will be given as the error message. | |
""" | |
if not protocol: | |
protocol = default | |
if protocol not in registry: | |
if protocol not in known_implementations: | |
raise ValueError(f"Protocol not known: {protocol}") | |
bit = known_implementations[protocol] | |
try: | |
register_implementation(protocol, _import_class(bit["class"])) | |
except ImportError as e: | |
raise ImportError(bit["err"]) from e | |
cls = registry[protocol] | |
if getattr(cls, "protocol", None) in ("abstract", None): | |
cls.protocol = protocol | |
return cls | |
s3_msg = """Your installed version of s3fs is very old and known to cause | |
severe performance issues, see also https://github.com/dask/dask/issues/10276 | |
To fix, you should specify a lower version bound on s3fs, or | |
update the current installation. | |
""" | |
def _import_class(cls, minv=None): | |
"""Take a string FQP and return the imported class or identifier | |
cls is of the form "package.module.klass" or "package.module:subobject.klass" | |
""" | |
if ":" in cls: | |
mod, name = cls.rsplit(":", 1) | |
s3 = mod == "s3fs" | |
mod = importlib.import_module(mod) | |
if s3 and mod.__version__.split(".") < ["0", "5"]: | |
warnings.warn(s3_msg) | |
for part in name.split("."): | |
mod = getattr(mod, part) | |
return mod | |
else: | |
mod, name = cls.rsplit(".", 1) | |
s3 = mod == "s3fs" | |
mod = importlib.import_module(mod) | |
if s3 and mod.__version__.split(".") < ["0", "5"]: | |
warnings.warn(s3_msg) | |
return getattr(mod, name) | |
def filesystem(protocol, **storage_options): | |
"""Instantiate filesystems for given protocol and arguments | |
``storage_options`` are specific to the protocol being chosen, and are | |
passed directly to the class. | |
""" | |
if protocol == "arrow_hdfs": | |
warnings.warn( | |
"The 'arrow_hdfs' protocol has been deprecated and will be " | |
"removed in the future. Specify it as 'hdfs'.", | |
DeprecationWarning, | |
) | |
cls = get_filesystem_class(protocol) | |
return cls(**storage_options) | |
def available_protocols(): | |
"""Return a list of the implemented protocols. | |
Note that any given protocol may require extra packages to be importable. | |
""" | |
return list(known_implementations) | |