Skip to content

Latest commit

 

History

History
270 lines (203 loc) · 11.9 KB

cephfs-snapshot-shallow-ro-vol.md

File metadata and controls

270 lines (203 loc) · 11.9 KB

Snapshots as shallow read-only volumes

CSI spec doesn't have a notion of "mounting a snapshot". Instead, the idiomatic way of accessing snapshot contents is first to create a volume populated with snapshot contents and then mount that volume to workloads.

CephFS exposes snapshots as special, read-only directories of a subvolume located in <subvolume>/.snap. cephfs-csi can already provision writable volumes with snapshots as their data source, where snapshot contents are cloned to the newly created volume. However, cloning a snapshot to volume is a very expensive operation in CephFS as the data needs to be fully copied. When the need is to only read snapshot contents, snapshot cloning is extremely inefficient and wasteful.

This proposal describes a way for cephfs-csi to expose CephFS snapshots as shallow, read-only volumes, without needing to clone the underlying snapshot data.

Use-cases

What's the point of such read-only volumes?

  • Restore snapshots selectively: users may want to traverse snapshots, restoring data to a writable volume more selectively instead of restoring the whole snapshot.

  • Volume backup: users can't backup a live volume, they first need to snapshot it. Once a snapshot is taken, it still can't be backed-up, as backup tools usually work with volumes (that are exposed as file-systems) and not snapshots (which might have backend-specific format). What this means is that in order to create a snapshot backup, users have to clone snapshot data twice:

    1. first time, when restoring the snapshot into a temporary volume from where the data will be read,
    2. and second time, when transferring that volume into some backup/archive storage (e.g. object store).

    The temporary backed-up volume will most likely be thrown away after the backup transfer is finished. That's a lot of wasted work for what we originally wanted to do! Having the ability to create volumes from snapshots cheaply would be a big improvement for this use case.

Alternatives

  • Snapshots are stored in <subvolume>/.snap. Users could simply visit this directory by themselves.

    .snap is CephFS-specific detail of how snapshots are exposed. Users / tools may not be aware of this special directory, or it may not fit their workflow. At the moment, the idiomatic way of accessing snapshot contents in CSI drivers is by creating a new volume and populating it with snapshot.

Design

Key points:

  • Volume source is a snapshot, volume access mode is *_READER_ONLY.
  • No actual new subvolumes are created in CephFS.
  • The resulting volume is a reference to the source subvolume snapshot. This reference would be stored in Volume.volume_context map. In order to reference a snapshot, we need subvolume name and snapshot name.
  • Mounting such volume means mounting the respective CephFS subvolume and exposing the snapshot to workloads.
  • Let's call a shallow read-only volume with a subvolume snapshot as its data source just a shallow volume from here on out for brevity.

Controller operations

Care must be taken when handling life-times of relevant storage resources. When a shallow volume is created, what would happen if:

  • Parent subvolume of the snapshot is removed while the shallow volume still exists?

    This shouldn't be a problem already. The parent volume has either snapshot-retention subvolume feature in which case its snapshots remain available, or if it doesn't have that feature, it will fail to be deleted because it still has snapshots associated to it.

  • Source snapshot from which the shallow volume originates is removed while that shallow volume still exists?

    We need to make sure this doesn't happen and some book-keeping is necessary. Ideally we could employ some kind of reference counting.

Reference counting for shallow volumes

As mentioned above, this is to protect shallow volumes, should their source snapshot be requested for deletion.

When creating a volume snapshot, a reference tracker (RT), represented by a RADOS object, would be created for that snapshot. It would store information required to track the references for the backing subvolume snapshot. Upon a CreateSnapshot call, the reference tracker (RT) would be initialized with a single reference record, where the CSI snapshot itself is the first reference to the backing snapshot. Each subsequent shallow volume creation would add a new reference record to the RT object. Each shallow volume deletion would remove that reference from the RT object. Calling DeleteSnapshot would remove the reference record that was previously added in CreateSnapshot.

The subvolume snapshot would be removed from the Ceph cluster only once the RT object holds no references. Note that this behavior would permit calling DeleteSnapshot even if it is still referenced by shallow volumes.

  • DeleteSnapshot:
  • RT holds no references or the RT object doesn't exist: delete the backing snapshot too.
  • RT holds at least one reference: keep the backing snapshot.
  • DeleteVolume:
  • RT holds no references: delete the backing snapshot too.
  • RT holds at least one reference: keep the backing snapshot.

To enable creating shallow volumes from snapshots that were provisioned by older versions of cephfs-csi (i.e. before this feature is introduced), CreateVolume for shallow volumes would also create an RT object in case it's missing. It would be initialized to two: the source snapshot and the newly created shallow volume.

Concurrent access to RT objects

RADOS API provides access to compound atomic read and write operations. These will be used to implement reference tracking functionality, protecting modifications of reference records.

CreateVolume

A read-only volume with snapshot source would be created under these conditions:

  1. CreateVolumeRequest.volume_content_source is a snapshot,
  2. CreateVolumeRequest.volume_capabilities[*].access_mode is any of read-only volume access modes.
  3. Possibly other volume parameters in CreateVolumeRequest.parameters specific to shallow volumes.

CreateVolumeResponse.Volume.volume_context would then contain necessary information to identify the source subvolume / snapshot.

Things to look out for:

  • What's the volume size?

    It doesn't consume any space on the filesystem. Volume.capacity_bytes is allowed to contain zero. We could use that.

  • What should be the requested size when creating the volume (specified e.g. in PVC)?

    This one is tricky. CSI spec allows for CreateVolumeRequest.capacity_range.{required_bytes,limit_bytes} to be zero. On the other hand, PersistentVolumeClaim.spec.resources.requests.storage must be bigger than zero. cephfs-csi doesn't care about the requested size (the volume will be read-only, so it has no usable capacity) and would always set it to zero. This shouldn't case any problems for the time being, but still is something we should keep in mind.

CreateVolume and behavior when using volume as volume source (PVC-PVC clone):

New volume Source volume Behavior
shallow volume shallow volume Create a new reference to the parent snapshot of the source shallow volume.
regular volume shallow volume Equivalent for a request to create a regular volume with snapshot as its source.
shallow volume regular volume Such request doesn't make sense and CreateVolume should return an error.

DeleteVolume

Volume deletion is trivial.

CreateSnapshot

Snapshotting read-only volumes doesn't make sense in general, and should be rejected.

ControllerExpandVolume

Same thing as above. Expanding read-only volumes doesn't make sense in general, and should be rejected.

Node operations

Two cases need to be considered:

  • (a) Volume/snapshot provisioning is handled by cephfs-csi
  • (b) Volume/snapshot provisioning is handled externally (e.g. pre-provisioned manually, or by OpenStack Manila, ...)

NodeStageVolume, NodeUnstageVolume

Here we're mounting the source subvolume onto the node. Subsequent volume publish calls then use bind mounts to expose the snapshot directory located in .snap/<SNAPSHOT DIRECTORY NAME>. Unfortunately, we cannot mount snapshots directly because they are not visible during mount time. We need to mount the whole subvolume first, and only then perform the binds to target paths.

For case (a)

Subvolume paths are normally retrieved by ceph fs subvolume info/getpath <VOLUME NAME> <SUBVOLUME NAME> <SUBVOLUMEGROUP NAME> , which outputs a path like so:

/volumes/<VOLUME NAME>/<SUBVOLUME NAME>/<UUID>

Snapshots are then accessible in:

  • /volumes/<VOLUME NAME>/<SUBVOLUME NAME>/.snap and
  • /volumes/<VOLUME NAME>/<SUBVOLUME NAME>/<UUID>/.snap.

/volumes/<VOLUME NAME>/<SUBVOLUME NAME>/<UUID> may be deleted if the source subvolume is deleted, but thanks to the snapshot-retention feature, snapshots in /volumes/<VOLUME NAME>/<SUBVOLUME NAME>/.snap will remain to be available.

The CephFS mount should therefore have its root set to the parent of what fs subvolume getpath returns, i.e. /volumes/<VOLUME NAME>/<SUBVOLUME NAME>. That way we will have snapshots available regardless of whether the subvolume itself still exists or not.

For case (b)

For cases where subvolumes are managed externally and not by cephfs-csi, we must assume that the cephx user we're given can access only /volumes/<VOLUME NAME>/<SUBVOLUME NAME>/<UUID> so users won't be able to benefit from snapshot retention. Users will need to be careful not to delete the parent subvolumes and snapshots while they are associated by these shallow RO volumes.

NodePublishVolume, NodeUnpublishVolume

Node publish is trivial. We bind staging path to target path as a read-only mount.

NodeGetVolumeStats

NodeGetVolumeStatsResponse.usage[*].available should be always zero.

Volume parameters, volume context

This section provides a discussion around determining what volume parameters and volume context parameters will be used to convey necessary information to the cephfs-csi driver in order to support shallow volumes.

Volume parameters CreateVolumeRequest.parameters:

  • Should be "shallow" the default mode for all CreateVolume calls that have (a) snapshot as data source and (b) read-only volume access mode? If not, a new volume parameter should be introduced: e.g isShallow: <bool>. On the other hand, does it even makes sense for users to want to create full copies of snapshots and still have them read-only?

Volume context Volume.volume_context:

  • Here we definitely need isShallow or similar. Without it we wouldn't be able to distinguish between a regular volume that just happens to have a read-only access mode, and a volume that references a snapshot.
  • Currently cephfs-csi recognizes subvolumePath for dynamically provisioned volumes and rootPath for pre-previsioned volumes. As mentioned in NodeStageVolume, NodeUnstageVolume section , snapshots cannot be mounted directly. How do we pass in path to the parent subvolume?
  • a) Path to the snapshot is passed in via subvolumePath / rootPath, e.g. /volumes/<VOLUME NAME>/<SUBVOLUME NAME>/<UUID>/.snap/<SNAPSHOT NAME>. From that we can derive path to the subvolume: it's the parent of .snap directory.
  • b) Similar to a), path to the snapshot is passed in via subvolumePath / rootPath, but instead of trying to derive the right path we introduce another volume context parameter containing path to the parent subvolume explicitly.
  • c) subvolumePath / rootPath contains path to the parent subvolume and we introduce another volume context parameter containing name of the snapshot. Path to the snapshot is then formed by appending /.snap/<SNAPSHOT NAME> to the subvolume path.