409 lines
13 KiB
Python
409 lines
13 KiB
Python
from __future__ import annotations
|
|
|
|
import inspect
|
|
import logging
|
|
import os
|
|
import shutil
|
|
import uuid
|
|
from typing import Optional
|
|
|
|
from .asyn import AsyncFileSystem, _run_coros_in_chunks, sync_wrapper
|
|
from .callbacks import DEFAULT_CALLBACK
|
|
from .core import filesystem, get_filesystem_class, split_protocol, url_to_fs
|
|
|
|
_generic_fs = {}
|
|
logger = logging.getLogger("fsspec.generic")
|
|
|
|
|
|
def set_generic_fs(protocol, **storage_options):
|
|
_generic_fs[protocol] = filesystem(protocol, **storage_options)
|
|
|
|
|
|
default_method = "default"
|
|
|
|
|
|
def _resolve_fs(url, method=None, protocol=None, storage_options=None):
|
|
"""Pick instance of backend FS"""
|
|
method = method or default_method
|
|
protocol = protocol or split_protocol(url)[0]
|
|
storage_options = storage_options or {}
|
|
if method == "default":
|
|
return filesystem(protocol)
|
|
if method == "generic":
|
|
return _generic_fs[protocol]
|
|
if method == "current":
|
|
cls = get_filesystem_class(protocol)
|
|
return cls.current()
|
|
if method == "options":
|
|
fs, _ = url_to_fs(url, **storage_options.get(protocol, {}))
|
|
return fs
|
|
raise ValueError(f"Unknown FS resolution method: {method}")
|
|
|
|
|
|
def rsync(
|
|
source,
|
|
destination,
|
|
delete_missing=False,
|
|
source_field="size",
|
|
dest_field="size",
|
|
update_cond="different",
|
|
inst_kwargs=None,
|
|
fs=None,
|
|
**kwargs,
|
|
):
|
|
"""Sync files between two directory trees
|
|
|
|
(experimental)
|
|
|
|
Parameters
|
|
----------
|
|
source: str
|
|
Root of the directory tree to take files from. This must be a directory, but
|
|
do not include any terminating "/" character
|
|
destination: str
|
|
Root path to copy into. The contents of this location should be
|
|
identical to the contents of ``source`` when done. This will be made a
|
|
directory, and the terminal "/" should not be included.
|
|
delete_missing: bool
|
|
If there are paths in the destination that don't exist in the
|
|
source and this is True, delete them. Otherwise, leave them alone.
|
|
source_field: str | callable
|
|
If ``update_field`` is "different", this is the key in the info
|
|
of source files to consider for difference. Maybe a function of the
|
|
info dict.
|
|
dest_field: str | callable
|
|
If ``update_field`` is "different", this is the key in the info
|
|
of destination files to consider for difference. May be a function of
|
|
the info dict.
|
|
update_cond: "different"|"always"|"never"
|
|
If "always", every file is copied, regardless of whether it exists in
|
|
the destination. If "never", files that exist in the destination are
|
|
not copied again. If "different" (default), only copy if the info
|
|
fields given by ``source_field`` and ``dest_field`` (usually "size")
|
|
are different. Other comparisons may be added in the future.
|
|
inst_kwargs: dict|None
|
|
If ``fs`` is None, use this set of keyword arguments to make a
|
|
GenericFileSystem instance
|
|
fs: GenericFileSystem|None
|
|
Instance to use if explicitly given. The instance defines how to
|
|
to make downstream file system instances from paths.
|
|
|
|
Returns
|
|
-------
|
|
dict of the copy operations that were performed, {source: destination}
|
|
"""
|
|
fs = fs or GenericFileSystem(**(inst_kwargs or {}))
|
|
source = fs._strip_protocol(source)
|
|
destination = fs._strip_protocol(destination)
|
|
allfiles = fs.find(source, withdirs=True, detail=True)
|
|
if not fs.isdir(source):
|
|
raise ValueError("Can only rsync on a directory")
|
|
otherfiles = fs.find(destination, withdirs=True, detail=True)
|
|
dirs = [
|
|
a
|
|
for a, v in allfiles.items()
|
|
if v["type"] == "directory" and a.replace(source, destination) not in otherfiles
|
|
]
|
|
logger.debug(f"{len(dirs)} directories to create")
|
|
if dirs:
|
|
fs.make_many_dirs(
|
|
[dirn.replace(source, destination) for dirn in dirs], exist_ok=True
|
|
)
|
|
allfiles = {a: v for a, v in allfiles.items() if v["type"] == "file"}
|
|
logger.debug(f"{len(allfiles)} files to consider for copy")
|
|
to_delete = [
|
|
o
|
|
for o, v in otherfiles.items()
|
|
if o.replace(destination, source) not in allfiles and v["type"] == "file"
|
|
]
|
|
for k, v in allfiles.copy().items():
|
|
otherfile = k.replace(source, destination)
|
|
if otherfile in otherfiles:
|
|
if update_cond == "always":
|
|
allfiles[k] = otherfile
|
|
elif update_cond == "different":
|
|
inf1 = source_field(v) if callable(source_field) else v[source_field]
|
|
v2 = otherfiles[otherfile]
|
|
inf2 = dest_field(v2) if callable(dest_field) else v2[dest_field]
|
|
if inf1 != inf2:
|
|
# details mismatch, make copy
|
|
allfiles[k] = otherfile
|
|
else:
|
|
# details match, don't copy
|
|
allfiles.pop(k)
|
|
else:
|
|
# file not in target yet
|
|
allfiles[k] = otherfile
|
|
logger.debug(f"{len(allfiles)} files to copy")
|
|
if allfiles:
|
|
source_files, target_files = zip(*allfiles.items())
|
|
fs.cp(source_files, target_files, **kwargs)
|
|
logger.debug(f"{len(to_delete)} files to delete")
|
|
if delete_missing:
|
|
fs.rm(to_delete)
|
|
return allfiles
|
|
|
|
|
|
class GenericFileSystem(AsyncFileSystem):
|
|
"""Wrapper over all other FS types
|
|
|
|
<experimental!>
|
|
|
|
This implementation is a single unified interface to be able to run FS operations
|
|
over generic URLs, and dispatch to the specific implementations using the URL
|
|
protocol prefix.
|
|
|
|
Note: instances of this FS are always async, even if you never use it with any async
|
|
backend.
|
|
"""
|
|
|
|
protocol = "generic" # there is no real reason to ever use a protocol with this FS
|
|
|
|
def __init__(self, default_method="default", **kwargs):
|
|
"""
|
|
|
|
Parameters
|
|
----------
|
|
default_method: str (optional)
|
|
Defines how to configure backend FS instances. Options are:
|
|
- "default": instantiate like FSClass(), with no
|
|
extra arguments; this is the default instance of that FS, and can be
|
|
configured via the config system
|
|
- "generic": takes instances from the `_generic_fs` dict in this module,
|
|
which you must populate before use. Keys are by protocol
|
|
- "current": takes the most recently instantiated version of each FS
|
|
"""
|
|
self.method = default_method
|
|
super().__init__(**kwargs)
|
|
|
|
def _parent(self, path):
|
|
fs = _resolve_fs(path, self.method)
|
|
return fs.unstrip_protocol(fs._parent(path))
|
|
|
|
def _strip_protocol(self, path):
|
|
# normalization only
|
|
fs = _resolve_fs(path, self.method)
|
|
return fs.unstrip_protocol(fs._strip_protocol(path))
|
|
|
|
async def _find(self, path, maxdepth=None, withdirs=False, detail=False, **kwargs):
|
|
fs = _resolve_fs(path, self.method)
|
|
if fs.async_impl:
|
|
out = await fs._find(
|
|
path, maxdepth=maxdepth, withdirs=withdirs, detail=True, **kwargs
|
|
)
|
|
else:
|
|
out = fs.find(
|
|
path, maxdepth=maxdepth, withdirs=withdirs, detail=True, **kwargs
|
|
)
|
|
result = {}
|
|
for k, v in out.items():
|
|
name = fs.unstrip_protocol(k)
|
|
v["name"] = name
|
|
result[name] = v
|
|
if detail:
|
|
return result
|
|
return list(result)
|
|
|
|
async def _info(self, url, **kwargs):
|
|
fs = _resolve_fs(url, self.method)
|
|
if fs.async_impl:
|
|
out = await fs._info(url, **kwargs)
|
|
else:
|
|
out = fs.info(url, **kwargs)
|
|
out["name"] = fs.unstrip_protocol(out["name"])
|
|
return out
|
|
|
|
async def _ls(
|
|
self,
|
|
url,
|
|
detail=True,
|
|
**kwargs,
|
|
):
|
|
fs = _resolve_fs(url, self.method)
|
|
if fs.async_impl:
|
|
out = await fs._ls(url, detail=True, **kwargs)
|
|
else:
|
|
out = fs.ls(url, detail=True, **kwargs)
|
|
for o in out:
|
|
o["name"] = fs.unstrip_protocol(o["name"])
|
|
if detail:
|
|
return out
|
|
else:
|
|
return [o["name"] for o in out]
|
|
|
|
async def _cat_file(
|
|
self,
|
|
url,
|
|
**kwargs,
|
|
):
|
|
fs = _resolve_fs(url, self.method)
|
|
if fs.async_impl:
|
|
return await fs._cat_file(url, **kwargs)
|
|
else:
|
|
return fs.cat_file(url, **kwargs)
|
|
|
|
async def _pipe_file(
|
|
self,
|
|
path,
|
|
value,
|
|
**kwargs,
|
|
):
|
|
fs = _resolve_fs(path, self.method)
|
|
if fs.async_impl:
|
|
return await fs._pipe_file(path, value, **kwargs)
|
|
else:
|
|
return fs.pipe_file(path, value, **kwargs)
|
|
|
|
async def _rm(self, url, **kwargs):
|
|
urls = url
|
|
if isinstance(urls, str):
|
|
urls = [urls]
|
|
fs = _resolve_fs(urls[0], self.method)
|
|
if fs.async_impl:
|
|
await fs._rm(urls, **kwargs)
|
|
else:
|
|
fs.rm(url, **kwargs)
|
|
|
|
async def _makedirs(self, path, exist_ok=False):
|
|
logger.debug("Make dir %s", path)
|
|
fs = _resolve_fs(path, self.method)
|
|
if fs.async_impl:
|
|
await fs._makedirs(path, exist_ok=exist_ok)
|
|
else:
|
|
fs.makedirs(path, exist_ok=exist_ok)
|
|
|
|
def rsync(self, source, destination, **kwargs):
|
|
"""Sync files between two directory trees
|
|
|
|
See `func:rsync` for more details.
|
|
"""
|
|
rsync(source, destination, fs=self, **kwargs)
|
|
|
|
async def _cp_file(
|
|
self,
|
|
url,
|
|
url2,
|
|
blocksize=2**20,
|
|
callback=DEFAULT_CALLBACK,
|
|
**kwargs,
|
|
):
|
|
fs = _resolve_fs(url, self.method)
|
|
fs2 = _resolve_fs(url2, self.method)
|
|
if fs is fs2:
|
|
# pure remote
|
|
if fs.async_impl:
|
|
return await fs._cp_file(url, url2, **kwargs)
|
|
else:
|
|
return fs.cp_file(url, url2, **kwargs)
|
|
kw = {"blocksize": 0, "cache_type": "none"}
|
|
try:
|
|
f1 = (
|
|
await fs.open_async(url, "rb")
|
|
if hasattr(fs, "open_async")
|
|
else fs.open(url, "rb", **kw)
|
|
)
|
|
callback.set_size(await maybe_await(f1.size))
|
|
f2 = (
|
|
await fs2.open_async(url2, "wb")
|
|
if hasattr(fs2, "open_async")
|
|
else fs2.open(url2, "wb", **kw)
|
|
)
|
|
while f1.size is None or f2.tell() < f1.size:
|
|
data = await maybe_await(f1.read(blocksize))
|
|
if f1.size is None and not data:
|
|
break
|
|
await maybe_await(f2.write(data))
|
|
callback.absolute_update(f2.tell())
|
|
finally:
|
|
try:
|
|
await maybe_await(f2.close())
|
|
await maybe_await(f1.close())
|
|
except NameError:
|
|
# fail while opening f1 or f2
|
|
pass
|
|
|
|
async def _make_many_dirs(self, urls, exist_ok=True):
|
|
fs = _resolve_fs(urls[0], self.method)
|
|
if fs.async_impl:
|
|
coros = [fs._makedirs(u, exist_ok=exist_ok) for u in urls]
|
|
await _run_coros_in_chunks(coros)
|
|
else:
|
|
for u in urls:
|
|
fs.makedirs(u, exist_ok=exist_ok)
|
|
|
|
make_many_dirs = sync_wrapper(_make_many_dirs)
|
|
|
|
async def _copy(
|
|
self,
|
|
path1: list[str],
|
|
path2: list[str],
|
|
recursive: bool = False,
|
|
on_error: str = "ignore",
|
|
maxdepth: Optional[int] = None,
|
|
batch_size: Optional[int] = None,
|
|
tempdir: Optional[str] = None,
|
|
**kwargs,
|
|
):
|
|
if recursive:
|
|
raise NotImplementedError
|
|
fs = _resolve_fs(path1[0], self.method)
|
|
fs2 = _resolve_fs(path2[0], self.method)
|
|
# not expanding paths atm., assume call is from rsync()
|
|
if fs is fs2:
|
|
# pure remote
|
|
if fs.async_impl:
|
|
return await fs._copy(path1, path2, **kwargs)
|
|
else:
|
|
return fs.copy(path1, path2, **kwargs)
|
|
await copy_file_op(
|
|
fs, path1, fs2, path2, tempdir, batch_size, on_error=on_error
|
|
)
|
|
|
|
|
|
async def copy_file_op(
|
|
fs1, url1, fs2, url2, tempdir=None, batch_size=20, on_error="ignore"
|
|
):
|
|
import tempfile
|
|
|
|
tempdir = tempdir or tempfile.mkdtemp()
|
|
try:
|
|
coros = [
|
|
_copy_file_op(
|
|
fs1,
|
|
u1,
|
|
fs2,
|
|
u2,
|
|
os.path.join(tempdir, uuid.uuid4().hex),
|
|
on_error=on_error,
|
|
)
|
|
for u1, u2 in zip(url1, url2)
|
|
]
|
|
await _run_coros_in_chunks(coros, batch_size=batch_size)
|
|
finally:
|
|
shutil.rmtree(tempdir)
|
|
|
|
|
|
async def _copy_file_op(fs1, url1, fs2, url2, local, on_error="ignore"):
|
|
ex = () if on_error == "raise" else Exception
|
|
logger.debug("Copy %s -> %s", url1, url2)
|
|
try:
|
|
if fs1.async_impl:
|
|
await fs1._get_file(url1, local)
|
|
else:
|
|
fs1.get_file(url1, local)
|
|
if fs2.async_impl:
|
|
await fs2._put_file(local, url2)
|
|
else:
|
|
fs2.put_file(local, url2)
|
|
os.unlink(local)
|
|
logger.debug("Copy %s -> %s; done", url1, url2)
|
|
except ex as e:
|
|
logger.debug("ignoring cp exception for %s: %s", url1, e)
|
|
|
|
|
|
async def maybe_await(cor):
|
|
if inspect.iscoroutine(cor):
|
|
return await cor
|
|
else:
|
|
return cor
|