diff --git a/pbs-client/src/pxar/mod.rs b/pbs-client/src/pxar/mod.rs index 66150178..c55c942b 100644 --- a/pbs-client/src/pxar/mod.rs +++ b/pbs-client/src/pxar/mod.rs @@ -32,7 +32,7 @@ //! //! * `FILENAME` -- name of the first directory entry (strictly ordered!) //! * `` -- serialization of the first directory entry's metadata and contents, -//! following the exact same archive format +//! following the exact same archive format //! * `FILENAME` -- name of the second directory entry (strictly ordered!) //! * `` -- serialization of the second directory entry //! * ... diff --git a/pbs-config/src/acl.rs b/pbs-config/src/acl.rs index a06b918a..d8138078 100644 --- a/pbs-config/src/acl.rs +++ b/pbs-config/src/acl.rs @@ -646,8 +646,8 @@ impl AclTree { /// - iterate over all intermediate nodes along `path` and collect roles with `propagate` set /// - get all (propagating and non-propagating) roles for last component of path /// - more specific role maps replace less specific role maps - /// -- user/token is more specific than group at each level - /// -- roles lower in the tree are more specific than those higher up along the path + /// -- user/token is more specific than group at each level + /// -- roles lower in the tree are more specific than those higher up along the path pub fn roles(&self, auth_id: &Authid, path: &[&str]) -> HashMap { let mut node = &self.root; let mut role_map = node.extract_roles(auth_id, path.is_empty()); diff --git a/src/server/pull.rs b/src/server/pull.rs index 361ed068..516abfe5 100644 --- a/src/server/pull.rs +++ b/src/server/pull.rs @@ -246,8 +246,8 @@ fn verify_archive(info: &FileInfo, csum: &[u8; 32], size: u64) -> Result<(), Err /// /// Pulling an archive consists of the following steps: /// - Load archive file into tmp file -/// -- Load file into tmp file -/// -- Verify tmp file checksum +/// -- Load file into tmp file +/// -- Verify tmp file checksum /// - if archive is an index, pull referenced chunks /// - Rename tmp file into real path async fn pull_single_archive<'a>( @@ -328,10 +328,10 @@ async fn pull_single_archive<'a>( /// /// Pulling a snapshot consists of the following steps: /// - (Re)download the manifest -/// -- if it matches and is not corrupt, only download log and treat snapshot as already synced +/// -- if it matches and is not corrupt, only download log and treat snapshot as already synced /// - Iterate over referenced files -/// -- if file already exists, verify contents -/// -- if not, pull it from the remote +/// -- if file already exists, verify contents +/// -- if not, pull it from the remote /// - Download log if not already existing async fn pull_snapshot<'a>( reader: Arc, @@ -495,7 +495,7 @@ async fn pull_snapshot_from<'a>( /// - Sort by snapshot time /// - Get last snapshot timestamp on local datastore /// - Iterate over list of snapshots -/// -- pull snapshot, unless it's not finished yet or older than last local snapshot +/// -- pull snapshot, unless it's not finished yet or older than last local snapshot /// - (remove_vanished) list all local snapshots, remove those that don't exist on remote /// /// Backwards-compat: if `source_namespace` is [None], only the group type and ID will be sent to the @@ -760,8 +760,8 @@ fn check_and_remove_vanished_ns( /// Pulling a store consists of the following steps: /// - Query list of namespaces on the remote /// - Iterate list -/// -- create sub-NS if needed (and allowed) -/// -- attempt to pull each NS in turn +/// -- create sub-NS if needed (and allowed) +/// -- attempt to pull each NS in turn /// - (remove_vanished && max_depth > 0) remove sub-NS which are not or no longer available on the remote /// /// Backwards compat: if the remote namespace is `/` and recursion is disabled, no namespace is