diff --git a/swh/fuse/cache.py b/swh/fuse/cache.py
index f99626b80ac2aaf37c7153bddc53c3482301e6a7..8c6986bb7a1b2164613865ccd656bce59352ed46 100644
--- a/swh/fuse/cache.py
+++ b/swh/fuse/cache.py
@@ -80,7 +80,7 @@ class FuseCache:
         await self.history.__aexit__()
 
     async def get_cached_swhids(self) -> AsyncGenerator[CoreSWHID, None]:
-        """ Return a list of all previously cached SWHID """
+        """Return a list of all previously cached SWHID"""
 
         # Use the metadata db since it should always contain all accessed SWHIDs
         metadata_cursor = await self.metadata.conn.execute(
@@ -91,7 +91,7 @@ class FuseCache:
             yield CoreSWHID.from_string(raw_swhid[0])
 
     async def get_cached_visits(self) -> AsyncGenerator[str, None]:
-        """ Return a list of all previously cached visit URL """
+        """Return a list of all previously cached visit URL"""
 
         cursor = await self.metadata.conn.execute("select url from visits_cache")
         urls = await cursor.fetchall()
@@ -100,7 +100,7 @@ class FuseCache:
 
 
 class AbstractCache(ABC):
-    """ Abstract cache implementation to share common behavior between cache types """
+    """Abstract cache implementation to share common behavior between cache types"""
 
     DB_SCHEMA: str = ""
     conf: Dict[str, Any]
@@ -129,11 +129,11 @@ class AbstractCache(ABC):
 
 
 class MetadataCache(AbstractCache):
-    """ The metadata cache map each artifact to the complete metadata of the
+    """The metadata cache map each artifact to the complete metadata of the
     referenced object. This is analogous to what is available in
     `archive/<SWHID>.json` file (and generally used as data source for returning
     the content of those files). Artifacts are identified using their SWHIDs, or
-    in the case of origin visits, using their URLs. """
+    in the case of origin visits, using their URLs."""
 
     DB_SCHEMA = """
         create table if not exists metadata_cache (
@@ -166,7 +166,8 @@ class MetadataCache(AbstractCache):
 
     async def get_visits(self, url_encoded: str) -> Optional[List[Dict[str, Any]]]:
         cursor = await self.conn.execute(
-            "select metadata, itime from visits_cache where url=?", (url_encoded,),
+            "select metadata, itime from visits_cache where url=?",
+            (url_encoded,),
         )
         cache = await cursor.fetchone()
         if cache:
@@ -206,19 +207,20 @@ class MetadataCache(AbstractCache):
 
     async def remove(self, swhid: CoreSWHID) -> None:
         await self.conn.execute(
-            "delete from metadata_cache where swhid=?", (str(swhid),),
+            "delete from metadata_cache where swhid=?",
+            (str(swhid),),
         )
         await self.conn.commit()
 
 
 class BlobCache(AbstractCache):
-    """ The blob cache map SWHIDs of type `cnt` to the bytes of their archived
+    """The blob cache map SWHIDs of type `cnt` to the bytes of their archived
     content.
 
     The blob cache entry for a given content object is populated, at the latest,
     the first time the object is `read()`-d. It might be populated earlier on
     due to prefetching, e.g., when a directory pointing to the given content is
-    listed for the first time. """
+    listed for the first time."""
 
     DB_SCHEMA = """
         create table if not exists blob_cache (
@@ -246,18 +248,19 @@ class BlobCache(AbstractCache):
 
     async def remove(self, swhid: CoreSWHID) -> None:
         await self.conn.execute(
-            "delete from blob_cache where swhid=?", (str(swhid),),
+            "delete from blob_cache where swhid=?",
+            (str(swhid),),
         )
         await self.conn.commit()
 
 
 class HistoryCache(AbstractCache):
-    """ The history cache map SWHIDs of type `rev` to a list of `rev` SWHIDs
+    """The history cache map SWHIDs of type `rev` to a list of `rev` SWHIDs
     corresponding to all its revision ancestors, sorted in reverse topological
     order. As the parents cache, the history cache is lazily populated and can
     be prefetched. To efficiently store the ancestor lists, the history cache
     represents ancestors as graph edges (a pair of two SWHID nodes), meaning the
-    history cache is shared amongst all revisions parents. """
+    history cache is shared amongst all revisions parents."""
 
     DB_SCHEMA = """
         create table if not exists history_graph (
@@ -282,7 +285,10 @@ class HistoryCache(AbstractCache):
     """
 
     async def get(self, swhid: CoreSWHID) -> Optional[List[CoreSWHID]]:
-        cursor = await self.conn.execute(self.HISTORY_REC_QUERY, (str(swhid),),)
+        cursor = await self.conn.execute(
+            self.HISTORY_REC_QUERY,
+            (str(swhid),),
+        )
         cache = await cursor.fetchall()
         if not cache:
             return None
@@ -330,7 +336,7 @@ class HistoryCache(AbstractCache):
 
 
 class DirEntryCache:
-    """ The direntry cache map inode representing directories to the entries
+    """The direntry cache map inode representing directories to the entries
     they contain. Each entry comes with its name as well as file attributes
     (i.e., all its needed to perform a detailed directory listing).
 
@@ -341,7 +347,7 @@ class DirEntryCache:
     content of the directory is listed. More aggressive prefetching might
     happen. For instance, when first opening a dir a recursive listing of it can
     be retrieved from the remote backend and used to recursively populate the
-    direntry cache for all (transitive) sub-directories. """
+    direntry cache for all (transitive) sub-directories."""
 
     @dataclass
     class LRU(OrderedDict):
@@ -385,7 +391,7 @@ class DirEntryCache:
         if unit == "%":
             max_ram = int(num * virtual_memory().available / 100)
         else:
-            units = {"B": 1, "KB": 10 ** 3, "MB": 10 ** 6, "GB": 10 ** 9}
+            units = {"B": 1, "KB": 10**3, "MB": 10**6, "GB": 10**9}
             max_ram = int(float(num) * units[unit])
 
         self.lru_cache = self.LRU(max_ram)
diff --git a/swh/fuse/cli.py b/swh/fuse/cli.py
index 62474aab06b6727ee286fa50bc57f82c47ec6e0e..814d419c242d31fad3dc033de3981652d41026bd 100644
--- a/swh/fuse/cli.py
+++ b/swh/fuse/cli.py
@@ -161,7 +161,10 @@ def mount(ctx, swhids, path, foreground):
                         },
                     },
                     "loggers": {
-                        LOGGER_NAME: {"level": log_level, "handlers": ["syslog"],},
+                        LOGGER_NAME: {
+                            "level": log_level,
+                            "handlers": ["syslog"],
+                        },
                     },
                 }
             )
@@ -202,9 +205,7 @@ def umount(ctx, path):
 @fuse.command()
 @click.pass_context
 def clean(ctx):
-    """Clean on-disk cache(s).
-
-    """
+    """Clean on-disk cache(s)."""
 
     def rm_cache(conf, cache_name):
         try:
diff --git a/swh/fuse/fs/artifact.py b/swh/fuse/fs/artifact.py
index 2e8614e86ebf3a09d76947d60101f4e535374e83..8f5ad96872ca8a50c389cf8d3338706823af41c5 100644
--- a/swh/fuse/fs/artifact.py
+++ b/swh/fuse/fs/artifact.py
@@ -28,7 +28,7 @@ SWHID_REGEXP = r"swh:1:(cnt|dir|rel|rev|snp):[0-9a-f]{40}"
 
 @dataclass
 class Content(FuseFileEntry):
-    """ Software Heritage content artifact.
+    """Software Heritage content artifact.
 
     Content leaves (AKA blobs) are represented on disks as regular files,
     containing the corresponding bytes, as archived.
@@ -36,7 +36,7 @@ class Content(FuseFileEntry):
     Note that permissions are associated to blobs only in the context of
     directories. Hence, when accessing blobs from the top-level `archive/`
     directory, the permissions of the `archive/SWHID` file will be arbitrary and
-    not meaningful (e.g., `0x644`). """
+    not meaningful (e.g., `0x644`)."""
 
     swhid: CoreSWHID
     prefetch: Any = None
@@ -57,7 +57,7 @@ class Content(FuseFileEntry):
 
 @dataclass
 class Directory(FuseDirEntry):
-    """ Software Heritage directory artifact.
+    """Software Heritage directory artifact.
 
     Directory nodes are represented as directories on the file-system,
     containing one entry for each entry of the archived directory. Entry names
@@ -66,7 +66,7 @@ class Directory(FuseDirEntry):
 
     Note that the FUSE mount is read-only, no matter what the permissions say.
     So it is possible that, in the context of a directory, a file is presented
-    as writable, whereas actually writing to it will fail with `EPERM`. """
+    as writable, whereas actually writing to it will fail with `EPERM`."""
 
     swhid: CoreSWHID
 
@@ -93,7 +93,9 @@ class Directory(FuseDirEntry):
                     pass  # Ignore error and create a (broken) symlink anyway
 
                 yield self.create_child(
-                    FuseSymlinkEntry, name=name, target=target,
+                    FuseSymlinkEntry,
+                    name=name,
+                    target=target,
                 )
             # 2. Regular file
             elif swhid.object_type == ObjectType.CONTENT:
@@ -109,7 +111,10 @@ class Directory(FuseDirEntry):
             # 3. Regular directory
             elif swhid.object_type == ObjectType.DIRECTORY:
                 yield self.create_child(
-                    Directory, name=name, mode=mode, swhid=swhid,
+                    Directory,
+                    name=name,
+                    mode=mode,
+                    swhid=swhid,
                 )
             # 4. Submodule
             elif swhid.object_type == ObjectType.REVISION:
@@ -131,7 +136,7 @@ class Directory(FuseDirEntry):
 
 @dataclass
 class Revision(FuseDirEntry):
-    """ Software Heritage revision artifact.
+    """Software Heritage revision artifact.
 
     Revision (AKA commit) nodes are represented on the file-system as
     directories with the following entries:
@@ -149,7 +154,7 @@ class Revision(FuseDirEntry):
       in reverse topological order. The history can be listed through
       `by-date/`, `by-hash/` or `by-page/` with each its own sharding policy.
     - `meta.json`: metadata for the current node, as a symlink pointing to the
-      relevant `archive/<SWHID>.json` file """
+      relevant `archive/<SWHID>.json` file"""
 
     swhid: CoreSWHID
 
@@ -179,7 +184,9 @@ class Revision(FuseDirEntry):
 
         if len(parents) >= 1:
             yield self.create_child(
-                FuseSymlinkEntry, name="parent", target="parents/1/",
+                FuseSymlinkEntry,
+                name="parent",
+                target="parents/1/",
             )
 
         yield self.create_child(
@@ -192,7 +199,7 @@ class Revision(FuseDirEntry):
 
 @dataclass
 class RevisionParents(FuseDirEntry):
-    """ Revision virtual `parents/` directory """
+    """Revision virtual `parents/` directory"""
 
     parents: List[CoreSWHID]
 
@@ -208,7 +215,7 @@ class RevisionParents(FuseDirEntry):
 
 @dataclass
 class RevisionHistory(FuseDirEntry):
-    """ Revision virtual `history/` directory """
+    """Revision virtual `history/` directory"""
 
     swhid: CoreSWHID
 
@@ -262,7 +269,7 @@ class RevisionHistory(FuseDirEntry):
 
 @dataclass
 class RevisionHistoryShardByDate(FuseDirEntry):
-    """ Revision virtual `history/by-date` sharded directory """
+    """Revision virtual `history/by-date` sharded directory"""
 
     history_swhid: CoreSWHID
     prefix: str = field(default="")
@@ -273,7 +280,7 @@ class RevisionHistoryShardByDate(FuseDirEntry):
 
     @dataclass
     class StatusFile(FuseFileEntry):
-        """ Temporary file used to indicate loading progress in by-date/ """
+        """Temporary file used to indicate loading progress in by-date/"""
 
         name: str = field(init=False, default=".status")
         mode: int = field(init=False, default=int(EntryMode.RDONLY_FILE))
@@ -345,7 +352,7 @@ class RevisionHistoryShardByDate(FuseDirEntry):
 
 @dataclass
 class RevisionHistoryShardByHash(FuseDirEntry):
-    """ Revision virtual `history/by-hash` sharded directory """
+    """Revision virtual `history/by-hash` sharded directory"""
 
     history_swhid: CoreSWHID
     prefix: str = field(default="")
@@ -383,7 +390,7 @@ class RevisionHistoryShardByHash(FuseDirEntry):
 
 @dataclass
 class RevisionHistoryShardByPage(FuseDirEntry):
-    """ Revision virtual `history/by-page` sharded directory """
+    """Revision virtual `history/by-page` sharded directory"""
 
     history_swhid: CoreSWHID
     prefix: Optional[int] = field(default=None)
@@ -421,7 +428,7 @@ class RevisionHistoryShardByPage(FuseDirEntry):
 
 @dataclass
 class Release(FuseDirEntry):
-    """ Software Heritage release artifact.
+    """Software Heritage release artifact.
 
     Release nodes are represented on the file-system as directories with the
     following entries:
@@ -432,7 +439,7 @@ class Release(FuseDirEntry):
       (transitively) resolves to a directory. When present it is a symlink
       pointing into `archive/` to the SWHID of the given directory
     - `meta.json`: metadata for the current node, as a symlink pointing to the
-      relevant `archive/<SWHID>.json` file """
+      relevant `archive/<SWHID>.json` file"""
 
     swhid: CoreSWHID
 
@@ -480,7 +487,7 @@ class Release(FuseDirEntry):
 
 @dataclass
 class ReleaseType(FuseFileEntry):
-    """ Release type virtual file """
+    """Release type virtual file"""
 
     target_type: ObjectType
 
@@ -490,13 +497,13 @@ class ReleaseType(FuseFileEntry):
 
 @dataclass
 class Snapshot(FuseDirEntry):
-    """ Software Heritage snapshot artifact.
+    """Software Heritage snapshot artifact.
 
     Snapshot nodes are represented on the file-system as recursive directories
     following the branch names structure. For example, a branch named
     ``refs/tags/v1.0`` will be represented as a ``refs`` directory containing a
     ``tags`` directory containing a ``v1.0`` symlink pointing to the branch
-    target SWHID. """
+    target SWHID."""
 
     swhid: CoreSWHID
     prefix: str = field(default="")
@@ -526,7 +533,9 @@ class Snapshot(FuseDirEntry):
                     target = f"{root_path}/archive/{target_raw}"
 
                 yield self.create_child(
-                    FuseSymlinkEntry, name=next_prefix, target=Path(target),
+                    FuseSymlinkEntry,
+                    name=next_prefix,
+                    target=Path(target),
                 )
             else:
                 subdirs.add(next_prefix)
@@ -543,7 +552,7 @@ class Snapshot(FuseDirEntry):
 
 @dataclass
 class Origin(FuseDirEntry):
-    """ Software Heritage origin artifact.
+    """Software Heritage origin artifact.
 
     Origin nodes are represented on the file-system as directories with one
     entry for each origin visit.
@@ -552,7 +561,7 @@ class Origin(FuseDirEntry):
     multiple visits occur the same day only the first one is kept). Each visit
     directory contains a `meta.json` with associated metadata for the origin
     node, and potentially a `snapshot` symlink pointing to the visit's snapshot
-    node. """
+    node."""
 
     DATE_FMT = "{year:04d}-{month:02d}-{day:02d}"
     ENTRIES_REGEXP = re.compile(r"^[0-9]{4}-[0-9]{2}-[0-9]{2}$")
@@ -573,13 +582,16 @@ class Origin(FuseDirEntry):
             else:
                 seen_date.add(name)
                 yield self.create_child(
-                    OriginVisit, name=name, mode=int(EntryMode.RDONLY_DIR), meta=visit,
+                    OriginVisit,
+                    name=name,
+                    mode=int(EntryMode.RDONLY_DIR),
+                    meta=visit,
                 )
 
 
 @dataclass
 class OriginVisit(FuseDirEntry):
-    """ Origin visit virtual directory """
+    """Origin visit virtual directory"""
 
     meta: Dict[str, Any]
 
diff --git a/swh/fuse/fs/entry.py b/swh/fuse/fs/entry.py
index b7688f23e124aa990a67ff5145b039443d1fe56c..4222aafd7fa775ca2e757a0fa1a6a1648a225bf1 100644
--- a/swh/fuse/fs/entry.py
+++ b/swh/fuse/fs/entry.py
@@ -17,7 +17,7 @@ if TYPE_CHECKING:  # avoid cyclic import
 
 
 class EntryMode(IntEnum):
-    """ Default entry mode and permissions for the FUSE.
+    """Default entry mode and permissions for the FUSE.
 
     The FUSE mount is always read-only, even if permissions contradict this
     statement (in a context of a directory, entries are listed with permissions
@@ -35,8 +35,7 @@ class EntryMode(IntEnum):
 
 @dataclass
 class FuseEntry:
-    """ Main wrapper class to manipulate virtual FUSE entries
-    """
+    """Main wrapper class to manipulate virtual FUSE entries"""
 
     name: str
     """entry filename"""
@@ -55,7 +54,7 @@ class FuseEntry:
         self.file_info_attrs["keep_cache"] = True
 
     async def size(self) -> int:
-        """ Return the size (in bytes) of an entry """
+        """Return the size (in bytes) of an entry"""
 
         raise NotImplementedError
 
@@ -71,10 +70,10 @@ class FuseEntry:
 
 @dataclass
 class FuseFileEntry(FuseEntry):
-    """ FUSE virtual file entry """
+    """FUSE virtual file entry"""
 
     async def get_content(self) -> bytes:
-        """ Return the content of a file entry """
+        """Return the content of a file entry"""
 
         raise NotImplementedError
 
@@ -84,7 +83,7 @@ class FuseFileEntry(FuseEntry):
 
 @dataclass
 class FuseDirEntry(FuseEntry):
-    """ FUSE virtual directory entry """
+    """FUSE virtual directory entry"""
 
     ENTRIES_REGEXP: Optional[Pattern] = field(init=False, default=None)
 
@@ -92,8 +91,8 @@ class FuseDirEntry(FuseEntry):
         return 0
 
     def validate_entry(self, name: str) -> bool:
-        """ Return true if the name matches the directory entries regular
-        expression, and false otherwise """
+        """Return true if the name matches the directory entries regular
+        expression, and false otherwise"""
 
         if self.ENTRIES_REGEXP:
             return bool(re.match(self.ENTRIES_REGEXP, name))
@@ -101,12 +100,12 @@ class FuseDirEntry(FuseEntry):
             return True
 
     async def compute_entries(self):
-        """ Return the child entries of a directory entry """
+        """Return the child entries of a directory entry"""
 
         raise NotImplementedError
 
     async def get_entries(self, offset: int = 0) -> AsyncIterator[FuseEntry]:
-        """ Return the child entries of a directory entry using direntry cache """
+        """Return the child entries of a directory entry using direntry cache"""
 
         cache = self.fuse.cache.direntry.get(self)
         if cache:
@@ -121,7 +120,7 @@ class FuseDirEntry(FuseEntry):
             yield entries[i]
 
     async def lookup(self, name: str) -> Optional[FuseEntry]:
-        """ Look up a FUSE entry by name """
+        """Look up a FUSE entry by name"""
 
         async for entry in self.get_entries():
             if entry.name == name:
@@ -131,8 +130,7 @@ class FuseDirEntry(FuseEntry):
 
 @dataclass
 class FuseSymlinkEntry(FuseEntry):
-    """ FUSE virtual symlink entry
-    """
+    """FUSE virtual symlink entry"""
 
     mode: int = field(init=False, default=int(EntryMode.RDONLY_LNK))
     target: Union[str, bytes, Path]
@@ -142,6 +140,6 @@ class FuseSymlinkEntry(FuseEntry):
         return len(str(self.target))
 
     def get_target(self) -> Union[str, bytes, Path]:
-        """ Return the path target of a symlink entry """
+        """Return the path target of a symlink entry"""
 
         return self.target
diff --git a/swh/fuse/fs/mountpoint.py b/swh/fuse/fs/mountpoint.py
index 9741b0ddd4c34eb5cac56dc1bc86939da137370e..5439544b2696e4eb98ab90db04ed196916d5e618 100644
--- a/swh/fuse/fs/mountpoint.py
+++ b/swh/fuse/fs/mountpoint.py
@@ -26,7 +26,7 @@ JSON_SUFFIX = ".json"
 
 @dataclass
 class Root(FuseDirEntry):
-    """ The FUSE mountpoint, consisting of the archive/ and origin/ directories """
+    """The FUSE mountpoint, consisting of the archive/ and origin/ directories"""
 
     name: str = field(init=False, default="")
     mode: int = field(init=False, default=int(EntryMode.RDONLY_DIR))
@@ -41,12 +41,12 @@ class Root(FuseDirEntry):
 
 @dataclass
 class ArchiveDir(FuseDirEntry):
-    """ The `archive/` virtual directory allows to mount any artifact on the fly
+    """The `archive/` virtual directory allows to mount any artifact on the fly
     using its SWHID as name. The associated metadata of the artifact from the
     Software Heritage Web API can also be accessed through the `SWHID.json` file
     (in case of pagination, the JSON file will contain a complete version with
     all pages merged together). Note: the archive directory cannot be listed
-    with ls, but entries in it can be accessed (e.g., using cat or cd). """
+    with ls, but entries in it can be accessed (e.g., using cat or cd)."""
 
     name: str = field(init=False, default="archive")
     mode: int = field(init=False, default=int(EntryMode.RDONLY_DIR))
@@ -87,8 +87,8 @@ class ArchiveDir(FuseDirEntry):
 
 @dataclass
 class MetaEntry(FuseFileEntry):
-    """ An entry for a `archive/<SWHID>.json` file, containing all the SWHID's
-    metadata from the Software Heritage archive. """
+    """An entry for a `archive/<SWHID>.json` file, containing all the SWHID's
+    metadata from the Software Heritage archive."""
 
     swhid: CoreSWHID
 
@@ -106,9 +106,9 @@ class MetaEntry(FuseFileEntry):
 
 @dataclass
 class OriginDir(FuseDirEntry):
-    """ The origin/ directory is lazily populated with one entry per accessed
+    """The origin/ directory is lazily populated with one entry per accessed
     origin URL (mangled to create a valid UNIX filename). The URL encoding is
-    done using the percent-encoding mechanism described in RFC 3986. """
+    done using the percent-encoding mechanism described in RFC 3986."""
 
     name: str = field(init=False, default="origin")
     mode: int = field(init=False, default=int(EntryMode.RDONLY_DIR))
@@ -117,7 +117,9 @@ class OriginDir(FuseDirEntry):
 
     def create_origin_child(self, url_encoded: str) -> FuseEntry:
         return super().create_child(
-            Origin, name=url_encoded, mode=int(EntryMode.RDONLY_DIR),
+            Origin,
+            name=url_encoded,
+            mode=int(EntryMode.RDONLY_DIR),
         )
 
     async def compute_entries(self) -> AsyncIterator[FuseEntry]:
@@ -140,7 +142,7 @@ class OriginDir(FuseDirEntry):
 
 @dataclass
 class CacheDir(FuseDirEntry):
-    """ The cache/ directory is an on-disk representation of locally cached
+    """The cache/ directory is an on-disk representation of locally cached
     objects and metadata. Via this directory you can browse cached data and
     selectively remove them from the cache, freeing disk space. (See `swh fs
     clean` in the {ref}`CLI <swh-fuse-cli>` to completely empty the cache). The
@@ -209,7 +211,7 @@ class CacheDir(FuseDirEntry):
 
 @dataclass
 class Readme(FuseFileEntry):
-    """ Top-level README to explain briefly what is SwhFS. """
+    """Top-level README to explain briefly what is SwhFS."""
 
     name: str = field(init=False, default="README")
     mode: int = field(init=False, default=int(EntryMode.RDONLY_FILE))
diff --git a/swh/fuse/fuse.py b/swh/fuse/fuse.py
index 8cedd1fded4f608e0f1877d6bcf3b02eb72ca8d9..207bd7e16aaa9c7cdcd699dbf17c93fa1b0dd67f 100644
--- a/swh/fuse/fuse.py
+++ b/swh/fuse/fuse.py
@@ -26,8 +26,8 @@ from swh.web.client.client import WebAPIClient
 
 
 class Fuse(pyfuse3.Operations):
-    """ Software Heritage Filesystem in Userspace (FUSE). Locally mount parts of
-    the archive and navigate it as a virtual file system. """
+    """Software Heritage Filesystem in Userspace (FUSE). Locally mount parts of
+    the archive and navigate it as a virtual file system."""
 
     def __init__(self, root_path: Path, cache: FuseCache, conf: Dict[str, Any]):
         super(Fuse, self).__init__()
@@ -54,7 +54,7 @@ class Fuse(pyfuse3.Operations):
         pass
 
     def _alloc_inode(self, entry: FuseEntry) -> int:
-        """ Return a unique inode integer for a given entry """
+        """Return a unique inode integer for a given entry"""
 
         inode = self._next_inode
         self._next_inode += 1
@@ -74,7 +74,7 @@ class Fuse(pyfuse3.Operations):
             pass
 
     def inode2entry(self, inode: int) -> FuseEntry:
-        """ Return the entry matching a given inode """
+        """Return the entry matching a given inode"""
 
         try:
             return self._inode2entry[inode]
@@ -82,7 +82,7 @@ class Fuse(pyfuse3.Operations):
             raise pyfuse3.FUSEError(errno.ENOENT)
 
     async def get_metadata(self, swhid: CoreSWHID) -> Any:
-        """ Retrieve metadata for a given SWHID using Software Heritage API """
+        """Retrieve metadata for a given SWHID using Software Heritage API"""
 
         cache = await self.cache.metadata.get(swhid)
         if cache:
@@ -101,8 +101,8 @@ class Fuse(pyfuse3.Operations):
             raise
 
     async def get_blob(self, swhid: CoreSWHID) -> bytes:
-        """ Retrieve the blob bytes for a given content SWHID using Software
-        Heritage API """
+        """Retrieve the blob bytes for a given content SWHID using Software
+        Heritage API"""
 
         if swhid.object_type != ObjectType.CONTENT:
             raise pyfuse3.FUSEError(errno.EINVAL)
@@ -127,7 +127,7 @@ class Fuse(pyfuse3.Operations):
             raise
 
     async def get_history(self, swhid: CoreSWHID) -> List[CoreSWHID]:
-        """ Retrieve a revision's history using Software Heritage Graph API """
+        """Retrieve a revision's history using Software Heritage Graph API"""
 
         if swhid.object_type != ObjectType.REVISION:
             raise pyfuse3.FUSEError(errno.EINVAL)
@@ -158,12 +158,14 @@ class Fuse(pyfuse3.Operations):
             return []
 
     async def get_visits(self, url_encoded: str) -> List[Dict[str, Any]]:
-        """ Retrieve origin visits given an encoded-URL using Software Heritage API """
+        """Retrieve origin visits given an encoded-URL using Software Heritage API"""
 
         cache = await self.cache.metadata.get_visits(url_encoded)
         if cache:
             self.logger.debug(
-                "Found %d visits for origin '%s' in cache", len(cache), url_encoded,
+                "Found %d visits for origin '%s' in cache",
+                len(cache),
+                url_encoded,
             )
             return cache
 
@@ -197,7 +199,7 @@ class Fuse(pyfuse3.Operations):
             raise
 
     async def get_attrs(self, entry: FuseEntry) -> pyfuse3.EntryAttributes:
-        """ Return entry attributes """
+        """Return entry attributes"""
 
         attrs = pyfuse3.EntryAttributes()
         attrs.st_size = 0
@@ -214,20 +216,20 @@ class Fuse(pyfuse3.Operations):
     async def getattr(
         self, inode: int, _ctx: pyfuse3.RequestContext
     ) -> pyfuse3.EntryAttributes:
-        """ Get attributes for a given inode """
+        """Get attributes for a given inode"""
 
         entry = self.inode2entry(inode)
         return await self.get_attrs(entry)
 
     async def opendir(self, inode: int, _ctx: pyfuse3.RequestContext) -> int:
-        """ Open a directory referred by a given inode """
+        """Open a directory referred by a given inode"""
 
         # Re-use inode as directory handle
         self.logger.debug("opendir(inode=%d)", inode)
         return inode
 
     async def readdir(self, fh: int, offset: int, token: pyfuse3.ReaddirToken) -> None:
-        """ Read entries in an open directory """
+        """Read entries in an open directory"""
 
         # opendir() uses inode as directory handle
         inode = fh
@@ -254,7 +256,7 @@ class Fuse(pyfuse3.Operations):
     async def open(
         self, inode: int, _flags: int, _ctx: pyfuse3.RequestContext
     ) -> pyfuse3.FileInfo:
-        """ Open an inode and return a unique file handle """
+        """Open an inode and return a unique file handle"""
 
         # Re-use inode as file handle
         self.logger.debug("open(inode=%d)", inode)
@@ -262,7 +264,7 @@ class Fuse(pyfuse3.Operations):
         return pyfuse3.FileInfo(fh=inode, **entry.file_info_attrs)
 
     async def read(self, fh: int, offset: int, length: int) -> bytes:
-        """ Read `length` bytes from file handle `fh` at position `offset` """
+        """Read `length` bytes from file handle `fh` at position `offset`"""
 
         # open() uses inode as file handle
         inode = fh
@@ -283,7 +285,7 @@ class Fuse(pyfuse3.Operations):
     async def lookup(
         self, parent_inode: int, name: str, _ctx: pyfuse3.RequestContext
     ) -> pyfuse3.EntryAttributes:
-        """ Look up a directory entry by name and get its attributes """
+        """Look up a directory entry by name and get its attributes"""
 
         name = os.fsdecode(name)
         parent_entry = self.inode2entry(parent_inode)
@@ -314,7 +316,7 @@ class Fuse(pyfuse3.Operations):
     async def unlink(
         self, parent_inode: int, name: str, _ctx: pyfuse3.RequestContext
     ) -> None:
-        """ Remove a file """
+        """Remove a file"""
 
         name = os.fsdecode(name)
         parent_entry = self.inode2entry(parent_inode)
@@ -333,7 +335,7 @@ class Fuse(pyfuse3.Operations):
 
 
 async def main(swhids: List[CoreSWHID], root_path: Path, conf: Dict[str, Any]) -> None:
-    """ swh-fuse CLI entry-point """
+    """swh-fuse CLI entry-point"""
 
     # Use pyfuse3 asyncio layer to match the rest of Software Heritage codebase
     pyfuse3_asyncio.enable()
diff --git a/swh/fuse/tests/conftest.py b/swh/fuse/tests/conftest.py
index 7e0c8f474cc5c14910aa653f5349bd9b78ae13c6..10cb70ec63a9191dbffd519c559c28101a709d72 100644
--- a/swh/fuse/tests/conftest.py
+++ b/swh/fuse/tests/conftest.py
@@ -39,7 +39,10 @@ def fuse_mntdir(web_api_mock):
     tmpdir = TemporaryDirectory(suffix=".swh-fuse-test")
 
     config = {
-        "cache": {"metadata": {"in-memory": True}, "blob": {"in-memory": True},},
+        "cache": {
+            "metadata": {"in-memory": True},
+            "blob": {"in-memory": True},
+        },
         "web-api": {"url": API_URL, "auth-token": None},
         "json-indent": None,
     }
diff --git a/swh/fuse/tests/data/config.py b/swh/fuse/tests/data/config.py
index 9f8af8dfacfb0bf86b80a786a5ecdba59277f8bb..c10e814f3871bd644f5eed69afe1389dbd1ad0de 100644
--- a/swh/fuse/tests/data/config.py
+++ b/swh/fuse/tests/data/config.py
@@ -46,10 +46,22 @@ ROOT_SNP_MASTER_BRANCH = "swh:1:rev:430a9fd4c797c50cea26157141b2408073b2ed91"
 FAKE_SNP_SPECIAL_CASES_SWHID = "swh:1:snp:0000000000000000000000000000000000000000"
 FAKE_SNP_SPECIAL_CASES = {
     # All possible target types
-    "mycnt": {"target_type": "content", "target": remove_swhid_prefix(REGULAR_FILE),},
-    "mydir": {"target_type": "directory", "target": remove_swhid_prefix(ROOT_DIR),},
-    "myrev": {"target_type": "revision", "target": remove_swhid_prefix(ROOT_REV),},
-    "myrel": {"target_type": "release", "target": remove_swhid_prefix(ROOT_REL),},
+    "mycnt": {
+        "target_type": "content",
+        "target": remove_swhid_prefix(REGULAR_FILE),
+    },
+    "mydir": {
+        "target_type": "directory",
+        "target": remove_swhid_prefix(ROOT_DIR),
+    },
+    "myrev": {
+        "target_type": "revision",
+        "target": remove_swhid_prefix(ROOT_REV),
+    },
+    "myrel": {
+        "target_type": "release",
+        "target": remove_swhid_prefix(ROOT_REL),
+    },
     "refs/heads/master": {
         "target_type": "revision",
         "target": remove_swhid_prefix(ROOT_SNP_MASTER_BRANCH),
diff --git a/swh/fuse/tests/data/gen-api-data.py b/swh/fuse/tests/data/gen-api-data.py
index 211b4add19805fc9fddc526b4d257c8a21af4923..0d468966f20f475b03509a18ad318bc084e292fa 100755
--- a/swh/fuse/tests/data/gen-api-data.py
+++ b/swh/fuse/tests/data/gen-api-data.py
@@ -124,7 +124,7 @@ swh:1:rev:1f3cff91f6762b0f47f41025b5e2c5ac942479ba swh:1:rev:bc286c7f2ceb5c3d2e0
 swh:1:rev:bc286c7f2ceb5c3d2e06ec72f78d28842f94ef65 swh:1:rev:f038f4d533f897a29f9422510d1b3f0caac97388
 swh:1:rev:f038f4d533f897a29f9422510d1b3f0caac97388 swh:1:rev:d6b7c96c3eb29b9244ece0c046d3f372ff432d04
 swh:1:rev:d6b7c96c3eb29b9244ece0c046d3f372ff432d04 swh:1:rev:c01efc669f09508b55eced32d3c88702578a7c3e
-"""  # NoQA: E501
+"""  # NoQA: B950
             MOCK_ARCHIVE[url] = history
 
             hist_nodes = set(
diff --git a/swh/fuse/tests/test_cli.py b/swh/fuse/tests/test_cli.py
index d8ee8e7e529033bb29f156e4c5780c7f854c43eb..c7c1f1edad5ea9ea20c721d58a7e12ff6ddd417d 100644
--- a/swh/fuse/tests/test_cli.py
+++ b/swh/fuse/tests/test_cli.py
@@ -30,7 +30,12 @@ def test_clean_command(tmp_path):
     assert fake_blob_db.exists()
 
     CliRunner().invoke(
-        cli.fuse, args=["--config-file", str(config_path), "clean",],
+        cli.fuse,
+        args=[
+            "--config-file",
+            str(config_path),
+            "clean",
+        ],
     )
 
     assert not fake_metadata_db.exists()