diff --git a/swh/fuse/cache.py b/swh/fuse/cache.py index f99626b80ac2aaf37c7153bddc53c3482301e6a7..8c6986bb7a1b2164613865ccd656bce59352ed46 100644 --- a/swh/fuse/cache.py +++ b/swh/fuse/cache.py @@ -80,7 +80,7 @@ class FuseCache: await self.history.__aexit__() async def get_cached_swhids(self) -> AsyncGenerator[CoreSWHID, None]: - """ Return a list of all previously cached SWHID """ + """Return a list of all previously cached SWHID""" # Use the metadata db since it should always contain all accessed SWHIDs metadata_cursor = await self.metadata.conn.execute( @@ -91,7 +91,7 @@ class FuseCache: yield CoreSWHID.from_string(raw_swhid[0]) async def get_cached_visits(self) -> AsyncGenerator[str, None]: - """ Return a list of all previously cached visit URL """ + """Return a list of all previously cached visit URL""" cursor = await self.metadata.conn.execute("select url from visits_cache") urls = await cursor.fetchall() @@ -100,7 +100,7 @@ class FuseCache: class AbstractCache(ABC): - """ Abstract cache implementation to share common behavior between cache types """ + """Abstract cache implementation to share common behavior between cache types""" DB_SCHEMA: str = "" conf: Dict[str, Any] @@ -129,11 +129,11 @@ class AbstractCache(ABC): class MetadataCache(AbstractCache): - """ The metadata cache map each artifact to the complete metadata of the + """The metadata cache map each artifact to the complete metadata of the referenced object. This is analogous to what is available in `archive/<SWHID>.json` file (and generally used as data source for returning the content of those files). Artifacts are identified using their SWHIDs, or - in the case of origin visits, using their URLs. """ + in the case of origin visits, using their URLs.""" DB_SCHEMA = """ create table if not exists metadata_cache ( @@ -166,7 +166,8 @@ class MetadataCache(AbstractCache): async def get_visits(self, url_encoded: str) -> Optional[List[Dict[str, Any]]]: cursor = await self.conn.execute( - "select metadata, itime from visits_cache where url=?", (url_encoded,), + "select metadata, itime from visits_cache where url=?", + (url_encoded,), ) cache = await cursor.fetchone() if cache: @@ -206,19 +207,20 @@ class MetadataCache(AbstractCache): async def remove(self, swhid: CoreSWHID) -> None: await self.conn.execute( - "delete from metadata_cache where swhid=?", (str(swhid),), + "delete from metadata_cache where swhid=?", + (str(swhid),), ) await self.conn.commit() class BlobCache(AbstractCache): - """ The blob cache map SWHIDs of type `cnt` to the bytes of their archived + """The blob cache map SWHIDs of type `cnt` to the bytes of their archived content. The blob cache entry for a given content object is populated, at the latest, the first time the object is `read()`-d. It might be populated earlier on due to prefetching, e.g., when a directory pointing to the given content is - listed for the first time. """ + listed for the first time.""" DB_SCHEMA = """ create table if not exists blob_cache ( @@ -246,18 +248,19 @@ class BlobCache(AbstractCache): async def remove(self, swhid: CoreSWHID) -> None: await self.conn.execute( - "delete from blob_cache where swhid=?", (str(swhid),), + "delete from blob_cache where swhid=?", + (str(swhid),), ) await self.conn.commit() class HistoryCache(AbstractCache): - """ The history cache map SWHIDs of type `rev` to a list of `rev` SWHIDs + """The history cache map SWHIDs of type `rev` to a list of `rev` SWHIDs corresponding to all its revision ancestors, sorted in reverse topological order. As the parents cache, the history cache is lazily populated and can be prefetched. To efficiently store the ancestor lists, the history cache represents ancestors as graph edges (a pair of two SWHID nodes), meaning the - history cache is shared amongst all revisions parents. """ + history cache is shared amongst all revisions parents.""" DB_SCHEMA = """ create table if not exists history_graph ( @@ -282,7 +285,10 @@ class HistoryCache(AbstractCache): """ async def get(self, swhid: CoreSWHID) -> Optional[List[CoreSWHID]]: - cursor = await self.conn.execute(self.HISTORY_REC_QUERY, (str(swhid),),) + cursor = await self.conn.execute( + self.HISTORY_REC_QUERY, + (str(swhid),), + ) cache = await cursor.fetchall() if not cache: return None @@ -330,7 +336,7 @@ class HistoryCache(AbstractCache): class DirEntryCache: - """ The direntry cache map inode representing directories to the entries + """The direntry cache map inode representing directories to the entries they contain. Each entry comes with its name as well as file attributes (i.e., all its needed to perform a detailed directory listing). @@ -341,7 +347,7 @@ class DirEntryCache: content of the directory is listed. More aggressive prefetching might happen. For instance, when first opening a dir a recursive listing of it can be retrieved from the remote backend and used to recursively populate the - direntry cache for all (transitive) sub-directories. """ + direntry cache for all (transitive) sub-directories.""" @dataclass class LRU(OrderedDict): @@ -385,7 +391,7 @@ class DirEntryCache: if unit == "%": max_ram = int(num * virtual_memory().available / 100) else: - units = {"B": 1, "KB": 10 ** 3, "MB": 10 ** 6, "GB": 10 ** 9} + units = {"B": 1, "KB": 10**3, "MB": 10**6, "GB": 10**9} max_ram = int(float(num) * units[unit]) self.lru_cache = self.LRU(max_ram) diff --git a/swh/fuse/cli.py b/swh/fuse/cli.py index 62474aab06b6727ee286fa50bc57f82c47ec6e0e..814d419c242d31fad3dc033de3981652d41026bd 100644 --- a/swh/fuse/cli.py +++ b/swh/fuse/cli.py @@ -161,7 +161,10 @@ def mount(ctx, swhids, path, foreground): }, }, "loggers": { - LOGGER_NAME: {"level": log_level, "handlers": ["syslog"],}, + LOGGER_NAME: { + "level": log_level, + "handlers": ["syslog"], + }, }, } ) @@ -202,9 +205,7 @@ def umount(ctx, path): @fuse.command() @click.pass_context def clean(ctx): - """Clean on-disk cache(s). - - """ + """Clean on-disk cache(s).""" def rm_cache(conf, cache_name): try: diff --git a/swh/fuse/fs/artifact.py b/swh/fuse/fs/artifact.py index 2e8614e86ebf3a09d76947d60101f4e535374e83..8f5ad96872ca8a50c389cf8d3338706823af41c5 100644 --- a/swh/fuse/fs/artifact.py +++ b/swh/fuse/fs/artifact.py @@ -28,7 +28,7 @@ SWHID_REGEXP = r"swh:1:(cnt|dir|rel|rev|snp):[0-9a-f]{40}" @dataclass class Content(FuseFileEntry): - """ Software Heritage content artifact. + """Software Heritage content artifact. Content leaves (AKA blobs) are represented on disks as regular files, containing the corresponding bytes, as archived. @@ -36,7 +36,7 @@ class Content(FuseFileEntry): Note that permissions are associated to blobs only in the context of directories. Hence, when accessing blobs from the top-level `archive/` directory, the permissions of the `archive/SWHID` file will be arbitrary and - not meaningful (e.g., `0x644`). """ + not meaningful (e.g., `0x644`).""" swhid: CoreSWHID prefetch: Any = None @@ -57,7 +57,7 @@ class Content(FuseFileEntry): @dataclass class Directory(FuseDirEntry): - """ Software Heritage directory artifact. + """Software Heritage directory artifact. Directory nodes are represented as directories on the file-system, containing one entry for each entry of the archived directory. Entry names @@ -66,7 +66,7 @@ class Directory(FuseDirEntry): Note that the FUSE mount is read-only, no matter what the permissions say. So it is possible that, in the context of a directory, a file is presented - as writable, whereas actually writing to it will fail with `EPERM`. """ + as writable, whereas actually writing to it will fail with `EPERM`.""" swhid: CoreSWHID @@ -93,7 +93,9 @@ class Directory(FuseDirEntry): pass # Ignore error and create a (broken) symlink anyway yield self.create_child( - FuseSymlinkEntry, name=name, target=target, + FuseSymlinkEntry, + name=name, + target=target, ) # 2. Regular file elif swhid.object_type == ObjectType.CONTENT: @@ -109,7 +111,10 @@ class Directory(FuseDirEntry): # 3. Regular directory elif swhid.object_type == ObjectType.DIRECTORY: yield self.create_child( - Directory, name=name, mode=mode, swhid=swhid, + Directory, + name=name, + mode=mode, + swhid=swhid, ) # 4. Submodule elif swhid.object_type == ObjectType.REVISION: @@ -131,7 +136,7 @@ class Directory(FuseDirEntry): @dataclass class Revision(FuseDirEntry): - """ Software Heritage revision artifact. + """Software Heritage revision artifact. Revision (AKA commit) nodes are represented on the file-system as directories with the following entries: @@ -149,7 +154,7 @@ class Revision(FuseDirEntry): in reverse topological order. The history can be listed through `by-date/`, `by-hash/` or `by-page/` with each its own sharding policy. - `meta.json`: metadata for the current node, as a symlink pointing to the - relevant `archive/<SWHID>.json` file """ + relevant `archive/<SWHID>.json` file""" swhid: CoreSWHID @@ -179,7 +184,9 @@ class Revision(FuseDirEntry): if len(parents) >= 1: yield self.create_child( - FuseSymlinkEntry, name="parent", target="parents/1/", + FuseSymlinkEntry, + name="parent", + target="parents/1/", ) yield self.create_child( @@ -192,7 +199,7 @@ class Revision(FuseDirEntry): @dataclass class RevisionParents(FuseDirEntry): - """ Revision virtual `parents/` directory """ + """Revision virtual `parents/` directory""" parents: List[CoreSWHID] @@ -208,7 +215,7 @@ class RevisionParents(FuseDirEntry): @dataclass class RevisionHistory(FuseDirEntry): - """ Revision virtual `history/` directory """ + """Revision virtual `history/` directory""" swhid: CoreSWHID @@ -262,7 +269,7 @@ class RevisionHistory(FuseDirEntry): @dataclass class RevisionHistoryShardByDate(FuseDirEntry): - """ Revision virtual `history/by-date` sharded directory """ + """Revision virtual `history/by-date` sharded directory""" history_swhid: CoreSWHID prefix: str = field(default="") @@ -273,7 +280,7 @@ class RevisionHistoryShardByDate(FuseDirEntry): @dataclass class StatusFile(FuseFileEntry): - """ Temporary file used to indicate loading progress in by-date/ """ + """Temporary file used to indicate loading progress in by-date/""" name: str = field(init=False, default=".status") mode: int = field(init=False, default=int(EntryMode.RDONLY_FILE)) @@ -345,7 +352,7 @@ class RevisionHistoryShardByDate(FuseDirEntry): @dataclass class RevisionHistoryShardByHash(FuseDirEntry): - """ Revision virtual `history/by-hash` sharded directory """ + """Revision virtual `history/by-hash` sharded directory""" history_swhid: CoreSWHID prefix: str = field(default="") @@ -383,7 +390,7 @@ class RevisionHistoryShardByHash(FuseDirEntry): @dataclass class RevisionHistoryShardByPage(FuseDirEntry): - """ Revision virtual `history/by-page` sharded directory """ + """Revision virtual `history/by-page` sharded directory""" history_swhid: CoreSWHID prefix: Optional[int] = field(default=None) @@ -421,7 +428,7 @@ class RevisionHistoryShardByPage(FuseDirEntry): @dataclass class Release(FuseDirEntry): - """ Software Heritage release artifact. + """Software Heritage release artifact. Release nodes are represented on the file-system as directories with the following entries: @@ -432,7 +439,7 @@ class Release(FuseDirEntry): (transitively) resolves to a directory. When present it is a symlink pointing into `archive/` to the SWHID of the given directory - `meta.json`: metadata for the current node, as a symlink pointing to the - relevant `archive/<SWHID>.json` file """ + relevant `archive/<SWHID>.json` file""" swhid: CoreSWHID @@ -480,7 +487,7 @@ class Release(FuseDirEntry): @dataclass class ReleaseType(FuseFileEntry): - """ Release type virtual file """ + """Release type virtual file""" target_type: ObjectType @@ -490,13 +497,13 @@ class ReleaseType(FuseFileEntry): @dataclass class Snapshot(FuseDirEntry): - """ Software Heritage snapshot artifact. + """Software Heritage snapshot artifact. Snapshot nodes are represented on the file-system as recursive directories following the branch names structure. For example, a branch named ``refs/tags/v1.0`` will be represented as a ``refs`` directory containing a ``tags`` directory containing a ``v1.0`` symlink pointing to the branch - target SWHID. """ + target SWHID.""" swhid: CoreSWHID prefix: str = field(default="") @@ -526,7 +533,9 @@ class Snapshot(FuseDirEntry): target = f"{root_path}/archive/{target_raw}" yield self.create_child( - FuseSymlinkEntry, name=next_prefix, target=Path(target), + FuseSymlinkEntry, + name=next_prefix, + target=Path(target), ) else: subdirs.add(next_prefix) @@ -543,7 +552,7 @@ class Snapshot(FuseDirEntry): @dataclass class Origin(FuseDirEntry): - """ Software Heritage origin artifact. + """Software Heritage origin artifact. Origin nodes are represented on the file-system as directories with one entry for each origin visit. @@ -552,7 +561,7 @@ class Origin(FuseDirEntry): multiple visits occur the same day only the first one is kept). Each visit directory contains a `meta.json` with associated metadata for the origin node, and potentially a `snapshot` symlink pointing to the visit's snapshot - node. """ + node.""" DATE_FMT = "{year:04d}-{month:02d}-{day:02d}" ENTRIES_REGEXP = re.compile(r"^[0-9]{4}-[0-9]{2}-[0-9]{2}$") @@ -573,13 +582,16 @@ class Origin(FuseDirEntry): else: seen_date.add(name) yield self.create_child( - OriginVisit, name=name, mode=int(EntryMode.RDONLY_DIR), meta=visit, + OriginVisit, + name=name, + mode=int(EntryMode.RDONLY_DIR), + meta=visit, ) @dataclass class OriginVisit(FuseDirEntry): - """ Origin visit virtual directory """ + """Origin visit virtual directory""" meta: Dict[str, Any] diff --git a/swh/fuse/fs/entry.py b/swh/fuse/fs/entry.py index b7688f23e124aa990a67ff5145b039443d1fe56c..4222aafd7fa775ca2e757a0fa1a6a1648a225bf1 100644 --- a/swh/fuse/fs/entry.py +++ b/swh/fuse/fs/entry.py @@ -17,7 +17,7 @@ if TYPE_CHECKING: # avoid cyclic import class EntryMode(IntEnum): - """ Default entry mode and permissions for the FUSE. + """Default entry mode and permissions for the FUSE. The FUSE mount is always read-only, even if permissions contradict this statement (in a context of a directory, entries are listed with permissions @@ -35,8 +35,7 @@ class EntryMode(IntEnum): @dataclass class FuseEntry: - """ Main wrapper class to manipulate virtual FUSE entries - """ + """Main wrapper class to manipulate virtual FUSE entries""" name: str """entry filename""" @@ -55,7 +54,7 @@ class FuseEntry: self.file_info_attrs["keep_cache"] = True async def size(self) -> int: - """ Return the size (in bytes) of an entry """ + """Return the size (in bytes) of an entry""" raise NotImplementedError @@ -71,10 +70,10 @@ class FuseEntry: @dataclass class FuseFileEntry(FuseEntry): - """ FUSE virtual file entry """ + """FUSE virtual file entry""" async def get_content(self) -> bytes: - """ Return the content of a file entry """ + """Return the content of a file entry""" raise NotImplementedError @@ -84,7 +83,7 @@ class FuseFileEntry(FuseEntry): @dataclass class FuseDirEntry(FuseEntry): - """ FUSE virtual directory entry """ + """FUSE virtual directory entry""" ENTRIES_REGEXP: Optional[Pattern] = field(init=False, default=None) @@ -92,8 +91,8 @@ class FuseDirEntry(FuseEntry): return 0 def validate_entry(self, name: str) -> bool: - """ Return true if the name matches the directory entries regular - expression, and false otherwise """ + """Return true if the name matches the directory entries regular + expression, and false otherwise""" if self.ENTRIES_REGEXP: return bool(re.match(self.ENTRIES_REGEXP, name)) @@ -101,12 +100,12 @@ class FuseDirEntry(FuseEntry): return True async def compute_entries(self): - """ Return the child entries of a directory entry """ + """Return the child entries of a directory entry""" raise NotImplementedError async def get_entries(self, offset: int = 0) -> AsyncIterator[FuseEntry]: - """ Return the child entries of a directory entry using direntry cache """ + """Return the child entries of a directory entry using direntry cache""" cache = self.fuse.cache.direntry.get(self) if cache: @@ -121,7 +120,7 @@ class FuseDirEntry(FuseEntry): yield entries[i] async def lookup(self, name: str) -> Optional[FuseEntry]: - """ Look up a FUSE entry by name """ + """Look up a FUSE entry by name""" async for entry in self.get_entries(): if entry.name == name: @@ -131,8 +130,7 @@ class FuseDirEntry(FuseEntry): @dataclass class FuseSymlinkEntry(FuseEntry): - """ FUSE virtual symlink entry - """ + """FUSE virtual symlink entry""" mode: int = field(init=False, default=int(EntryMode.RDONLY_LNK)) target: Union[str, bytes, Path] @@ -142,6 +140,6 @@ class FuseSymlinkEntry(FuseEntry): return len(str(self.target)) def get_target(self) -> Union[str, bytes, Path]: - """ Return the path target of a symlink entry """ + """Return the path target of a symlink entry""" return self.target diff --git a/swh/fuse/fs/mountpoint.py b/swh/fuse/fs/mountpoint.py index 9741b0ddd4c34eb5cac56dc1bc86939da137370e..5439544b2696e4eb98ab90db04ed196916d5e618 100644 --- a/swh/fuse/fs/mountpoint.py +++ b/swh/fuse/fs/mountpoint.py @@ -26,7 +26,7 @@ JSON_SUFFIX = ".json" @dataclass class Root(FuseDirEntry): - """ The FUSE mountpoint, consisting of the archive/ and origin/ directories """ + """The FUSE mountpoint, consisting of the archive/ and origin/ directories""" name: str = field(init=False, default="") mode: int = field(init=False, default=int(EntryMode.RDONLY_DIR)) @@ -41,12 +41,12 @@ class Root(FuseDirEntry): @dataclass class ArchiveDir(FuseDirEntry): - """ The `archive/` virtual directory allows to mount any artifact on the fly + """The `archive/` virtual directory allows to mount any artifact on the fly using its SWHID as name. The associated metadata of the artifact from the Software Heritage Web API can also be accessed through the `SWHID.json` file (in case of pagination, the JSON file will contain a complete version with all pages merged together). Note: the archive directory cannot be listed - with ls, but entries in it can be accessed (e.g., using cat or cd). """ + with ls, but entries in it can be accessed (e.g., using cat or cd).""" name: str = field(init=False, default="archive") mode: int = field(init=False, default=int(EntryMode.RDONLY_DIR)) @@ -87,8 +87,8 @@ class ArchiveDir(FuseDirEntry): @dataclass class MetaEntry(FuseFileEntry): - """ An entry for a `archive/<SWHID>.json` file, containing all the SWHID's - metadata from the Software Heritage archive. """ + """An entry for a `archive/<SWHID>.json` file, containing all the SWHID's + metadata from the Software Heritage archive.""" swhid: CoreSWHID @@ -106,9 +106,9 @@ class MetaEntry(FuseFileEntry): @dataclass class OriginDir(FuseDirEntry): - """ The origin/ directory is lazily populated with one entry per accessed + """The origin/ directory is lazily populated with one entry per accessed origin URL (mangled to create a valid UNIX filename). The URL encoding is - done using the percent-encoding mechanism described in RFC 3986. """ + done using the percent-encoding mechanism described in RFC 3986.""" name: str = field(init=False, default="origin") mode: int = field(init=False, default=int(EntryMode.RDONLY_DIR)) @@ -117,7 +117,9 @@ class OriginDir(FuseDirEntry): def create_origin_child(self, url_encoded: str) -> FuseEntry: return super().create_child( - Origin, name=url_encoded, mode=int(EntryMode.RDONLY_DIR), + Origin, + name=url_encoded, + mode=int(EntryMode.RDONLY_DIR), ) async def compute_entries(self) -> AsyncIterator[FuseEntry]: @@ -140,7 +142,7 @@ class OriginDir(FuseDirEntry): @dataclass class CacheDir(FuseDirEntry): - """ The cache/ directory is an on-disk representation of locally cached + """The cache/ directory is an on-disk representation of locally cached objects and metadata. Via this directory you can browse cached data and selectively remove them from the cache, freeing disk space. (See `swh fs clean` in the {ref}`CLI <swh-fuse-cli>` to completely empty the cache). The @@ -209,7 +211,7 @@ class CacheDir(FuseDirEntry): @dataclass class Readme(FuseFileEntry): - """ Top-level README to explain briefly what is SwhFS. """ + """Top-level README to explain briefly what is SwhFS.""" name: str = field(init=False, default="README") mode: int = field(init=False, default=int(EntryMode.RDONLY_FILE)) diff --git a/swh/fuse/fuse.py b/swh/fuse/fuse.py index 8cedd1fded4f608e0f1877d6bcf3b02eb72ca8d9..207bd7e16aaa9c7cdcd699dbf17c93fa1b0dd67f 100644 --- a/swh/fuse/fuse.py +++ b/swh/fuse/fuse.py @@ -26,8 +26,8 @@ from swh.web.client.client import WebAPIClient class Fuse(pyfuse3.Operations): - """ Software Heritage Filesystem in Userspace (FUSE). Locally mount parts of - the archive and navigate it as a virtual file system. """ + """Software Heritage Filesystem in Userspace (FUSE). Locally mount parts of + the archive and navigate it as a virtual file system.""" def __init__(self, root_path: Path, cache: FuseCache, conf: Dict[str, Any]): super(Fuse, self).__init__() @@ -54,7 +54,7 @@ class Fuse(pyfuse3.Operations): pass def _alloc_inode(self, entry: FuseEntry) -> int: - """ Return a unique inode integer for a given entry """ + """Return a unique inode integer for a given entry""" inode = self._next_inode self._next_inode += 1 @@ -74,7 +74,7 @@ class Fuse(pyfuse3.Operations): pass def inode2entry(self, inode: int) -> FuseEntry: - """ Return the entry matching a given inode """ + """Return the entry matching a given inode""" try: return self._inode2entry[inode] @@ -82,7 +82,7 @@ class Fuse(pyfuse3.Operations): raise pyfuse3.FUSEError(errno.ENOENT) async def get_metadata(self, swhid: CoreSWHID) -> Any: - """ Retrieve metadata for a given SWHID using Software Heritage API """ + """Retrieve metadata for a given SWHID using Software Heritage API""" cache = await self.cache.metadata.get(swhid) if cache: @@ -101,8 +101,8 @@ class Fuse(pyfuse3.Operations): raise async def get_blob(self, swhid: CoreSWHID) -> bytes: - """ Retrieve the blob bytes for a given content SWHID using Software - Heritage API """ + """Retrieve the blob bytes for a given content SWHID using Software + Heritage API""" if swhid.object_type != ObjectType.CONTENT: raise pyfuse3.FUSEError(errno.EINVAL) @@ -127,7 +127,7 @@ class Fuse(pyfuse3.Operations): raise async def get_history(self, swhid: CoreSWHID) -> List[CoreSWHID]: - """ Retrieve a revision's history using Software Heritage Graph API """ + """Retrieve a revision's history using Software Heritage Graph API""" if swhid.object_type != ObjectType.REVISION: raise pyfuse3.FUSEError(errno.EINVAL) @@ -158,12 +158,14 @@ class Fuse(pyfuse3.Operations): return [] async def get_visits(self, url_encoded: str) -> List[Dict[str, Any]]: - """ Retrieve origin visits given an encoded-URL using Software Heritage API """ + """Retrieve origin visits given an encoded-URL using Software Heritage API""" cache = await self.cache.metadata.get_visits(url_encoded) if cache: self.logger.debug( - "Found %d visits for origin '%s' in cache", len(cache), url_encoded, + "Found %d visits for origin '%s' in cache", + len(cache), + url_encoded, ) return cache @@ -197,7 +199,7 @@ class Fuse(pyfuse3.Operations): raise async def get_attrs(self, entry: FuseEntry) -> pyfuse3.EntryAttributes: - """ Return entry attributes """ + """Return entry attributes""" attrs = pyfuse3.EntryAttributes() attrs.st_size = 0 @@ -214,20 +216,20 @@ class Fuse(pyfuse3.Operations): async def getattr( self, inode: int, _ctx: pyfuse3.RequestContext ) -> pyfuse3.EntryAttributes: - """ Get attributes for a given inode """ + """Get attributes for a given inode""" entry = self.inode2entry(inode) return await self.get_attrs(entry) async def opendir(self, inode: int, _ctx: pyfuse3.RequestContext) -> int: - """ Open a directory referred by a given inode """ + """Open a directory referred by a given inode""" # Re-use inode as directory handle self.logger.debug("opendir(inode=%d)", inode) return inode async def readdir(self, fh: int, offset: int, token: pyfuse3.ReaddirToken) -> None: - """ Read entries in an open directory """ + """Read entries in an open directory""" # opendir() uses inode as directory handle inode = fh @@ -254,7 +256,7 @@ class Fuse(pyfuse3.Operations): async def open( self, inode: int, _flags: int, _ctx: pyfuse3.RequestContext ) -> pyfuse3.FileInfo: - """ Open an inode and return a unique file handle """ + """Open an inode and return a unique file handle""" # Re-use inode as file handle self.logger.debug("open(inode=%d)", inode) @@ -262,7 +264,7 @@ class Fuse(pyfuse3.Operations): return pyfuse3.FileInfo(fh=inode, **entry.file_info_attrs) async def read(self, fh: int, offset: int, length: int) -> bytes: - """ Read `length` bytes from file handle `fh` at position `offset` """ + """Read `length` bytes from file handle `fh` at position `offset`""" # open() uses inode as file handle inode = fh @@ -283,7 +285,7 @@ class Fuse(pyfuse3.Operations): async def lookup( self, parent_inode: int, name: str, _ctx: pyfuse3.RequestContext ) -> pyfuse3.EntryAttributes: - """ Look up a directory entry by name and get its attributes """ + """Look up a directory entry by name and get its attributes""" name = os.fsdecode(name) parent_entry = self.inode2entry(parent_inode) @@ -314,7 +316,7 @@ class Fuse(pyfuse3.Operations): async def unlink( self, parent_inode: int, name: str, _ctx: pyfuse3.RequestContext ) -> None: - """ Remove a file """ + """Remove a file""" name = os.fsdecode(name) parent_entry = self.inode2entry(parent_inode) @@ -333,7 +335,7 @@ class Fuse(pyfuse3.Operations): async def main(swhids: List[CoreSWHID], root_path: Path, conf: Dict[str, Any]) -> None: - """ swh-fuse CLI entry-point """ + """swh-fuse CLI entry-point""" # Use pyfuse3 asyncio layer to match the rest of Software Heritage codebase pyfuse3_asyncio.enable() diff --git a/swh/fuse/tests/conftest.py b/swh/fuse/tests/conftest.py index 7e0c8f474cc5c14910aa653f5349bd9b78ae13c6..10cb70ec63a9191dbffd519c559c28101a709d72 100644 --- a/swh/fuse/tests/conftest.py +++ b/swh/fuse/tests/conftest.py @@ -39,7 +39,10 @@ def fuse_mntdir(web_api_mock): tmpdir = TemporaryDirectory(suffix=".swh-fuse-test") config = { - "cache": {"metadata": {"in-memory": True}, "blob": {"in-memory": True},}, + "cache": { + "metadata": {"in-memory": True}, + "blob": {"in-memory": True}, + }, "web-api": {"url": API_URL, "auth-token": None}, "json-indent": None, } diff --git a/swh/fuse/tests/data/config.py b/swh/fuse/tests/data/config.py index 9f8af8dfacfb0bf86b80a786a5ecdba59277f8bb..c10e814f3871bd644f5eed69afe1389dbd1ad0de 100644 --- a/swh/fuse/tests/data/config.py +++ b/swh/fuse/tests/data/config.py @@ -46,10 +46,22 @@ ROOT_SNP_MASTER_BRANCH = "swh:1:rev:430a9fd4c797c50cea26157141b2408073b2ed91" FAKE_SNP_SPECIAL_CASES_SWHID = "swh:1:snp:0000000000000000000000000000000000000000" FAKE_SNP_SPECIAL_CASES = { # All possible target types - "mycnt": {"target_type": "content", "target": remove_swhid_prefix(REGULAR_FILE),}, - "mydir": {"target_type": "directory", "target": remove_swhid_prefix(ROOT_DIR),}, - "myrev": {"target_type": "revision", "target": remove_swhid_prefix(ROOT_REV),}, - "myrel": {"target_type": "release", "target": remove_swhid_prefix(ROOT_REL),}, + "mycnt": { + "target_type": "content", + "target": remove_swhid_prefix(REGULAR_FILE), + }, + "mydir": { + "target_type": "directory", + "target": remove_swhid_prefix(ROOT_DIR), + }, + "myrev": { + "target_type": "revision", + "target": remove_swhid_prefix(ROOT_REV), + }, + "myrel": { + "target_type": "release", + "target": remove_swhid_prefix(ROOT_REL), + }, "refs/heads/master": { "target_type": "revision", "target": remove_swhid_prefix(ROOT_SNP_MASTER_BRANCH), diff --git a/swh/fuse/tests/data/gen-api-data.py b/swh/fuse/tests/data/gen-api-data.py index 211b4add19805fc9fddc526b4d257c8a21af4923..0d468966f20f475b03509a18ad318bc084e292fa 100755 --- a/swh/fuse/tests/data/gen-api-data.py +++ b/swh/fuse/tests/data/gen-api-data.py @@ -124,7 +124,7 @@ swh:1:rev:1f3cff91f6762b0f47f41025b5e2c5ac942479ba swh:1:rev:bc286c7f2ceb5c3d2e0 swh:1:rev:bc286c7f2ceb5c3d2e06ec72f78d28842f94ef65 swh:1:rev:f038f4d533f897a29f9422510d1b3f0caac97388 swh:1:rev:f038f4d533f897a29f9422510d1b3f0caac97388 swh:1:rev:d6b7c96c3eb29b9244ece0c046d3f372ff432d04 swh:1:rev:d6b7c96c3eb29b9244ece0c046d3f372ff432d04 swh:1:rev:c01efc669f09508b55eced32d3c88702578a7c3e -""" # NoQA: E501 +""" # NoQA: B950 MOCK_ARCHIVE[url] = history hist_nodes = set( diff --git a/swh/fuse/tests/test_cli.py b/swh/fuse/tests/test_cli.py index d8ee8e7e529033bb29f156e4c5780c7f854c43eb..c7c1f1edad5ea9ea20c721d58a7e12ff6ddd417d 100644 --- a/swh/fuse/tests/test_cli.py +++ b/swh/fuse/tests/test_cli.py @@ -30,7 +30,12 @@ def test_clean_command(tmp_path): assert fake_blob_db.exists() CliRunner().invoke( - cli.fuse, args=["--config-file", str(config_path), "clean",], + cli.fuse, + args=[ + "--config-file", + str(config_path), + "clean", + ], ) assert not fake_metadata_db.exists()