[HDDS-10657] Overwriting an Ozone Key only if it has not changed. (accepted)

Authors: Stephen ODonnell
2024-04-05

 

Summary
A minimal design illustrating how to replace a key in Ozone only if it has not changes since it was read.

Ozone offers write semantics where the last writer to commit a key wins. Therefore multiple writers can concurrently write the same key, and which ever commits last will effectively overwrite all data that came before it.

As an extension of this, there is no “locking” on a key which is being replaced.

For any key, but especially a large key, it can take significant time to read and write it. There are scenarios where it would be desirable to replace a key in Ozone, but only if the key has not changed since it was read. With the absence of a lock, such an operation is not possible today.

As Things Stand

Internally, all Ozone keys have both an objectID and UpdateID which are stored in OM as part of the key metadata.

Each time something changes on the key, whether it is data or metadata, the updateID is changed. It comes from the ratis transactionID and is generally an increasing number.

When an existing key is over written, its existing metadata including the ObjectID and ACLs are mirrored onto the new key version. The only metadata which is replaced is any custom metadata stored on the key by the user. Upon commit, the updateID is also changed to the current Ratis transaction ID.

Writing a key in Ozone is a 3 step process:

  1. The key is opened via an Open Key request from the client to OM
  2. The client writes data to the data nodes
  3. The client commits the key to OM via a Commit Key call.

Note, that as things stand, it is possible to lose metadata updates (eg ACL changes) when a key is overwritten.

  1. If the key exists, then a new copy of the key is open for writing.
  2. While the new copy is open, another process updates the ACLs for the key
  3. On commit, the new ACLs are not copied to the new key as the new key made a copy of the existing metadata at the time the key was opened.

With the technique described in the next section, that problem is removed in this design, as the ACL update will change the updateID, and the key will not be committed.

Atomic Key Replacement

In relational database applications, records are often assigned an update counter similar to the updateID for a key in Ozone. The data record can be read and displayed on a UI to be edited, and then written back to the database. However another user could have made an edit to the same record in the mean time, and if the record is written back without any checks, those edits could be lost.

To combat this, “optimistic locking” is used. With Optimistic locking, no locks are actually involved. The client reads the data along with the update counter. When it attempts to write the data back, it validates the record has not change by including the updateID in the update statement, eg:

update customerDetails
set <columns = values>
where customerID = :b1
and updateCounter = :b2

If no records are updated, the application must display an error or reload the customer record to handle the problem.

In Ozone the same concept can be used to perform an atomic update of a key only if it has not changed since the key details were originally read.

To do this:

  1. The client reads the key details as usual. The key details can be extended to include the existing updateID as it is currently not passed to the client. This field already exists, but when exposed to the client it will be referred to as the key generation.
  2. The client can inspect the read key details and decide if it wants to replace the key.
  3. The client opens a new key for writing with the same key name as the original, passing the previously read generation in a new field. Call this new field expectedGeneration.
  4. On OM, it receives the openKey request as usual and detects the presence of the expectedGeneration field.
  5. On OM, it first ensures that a key is present with the given key name and having a updateID == expectedGeneration. If so, it opens the key and stored the details including the expectedGeneration in the openKeyTable. As things stand, the other existing key metadata copied from the original key is stored in the openKeyTable too.
  6. The client continues to write the data as usual. This can be the same data in a different format (eg Ratis to EC conversion), or new data in the key depending on the application’s needs.
  7. On commit key, the client does not need to send the expectedGeneration again, as the open key contains it.
  8. On OM, on commit key, it validates the key still exists with the given key name and its stored updateID is unchanged when compared with the expectedGeneration. If so the key is committed, otherwise an error is returned to the client.

Note that any change to a key will change the updateID. This is existing behaviour, and committing a rewritten key will also modify the updateID. Note this also offers protection against concurrent rewrites.

An optional enhancement for large keys, is that on each block allocation the expectedGeneration can be checked against the current key version to ensure it has not changed. This would allow the rewrite to fail early if a large multi block key is modified.

Alternative Proposal

  1. Pass the expected expectedGeneration to the rewrite API which passes it down to the relevant key stream, effectively saving it on the client
  2. Client attaches the expectedGeneration to the commit request to indicate a rewrite instead of a put
  3. OM checks the passed generation against the stored update ID and returns the corresponding success/fail result

The advantage of this alternative approach is that it does not require the expectedGeneration to be stored in the openKey table.

However the client code required to implement this appears more complex due to having different key commit logic for Ratis and EC and the parameter needing to be passed through many method calls.

PR #5524 illustrates this approach for the atomicKeyCreation feature which was added to S3.

The existing implementation for key creation stores various attributes (metadata, creation time, ACLs, ReplicationConfig) in the openKey table, so storing the expectedGeneration keeps with that convention, which is less confusing for future developers.

In terms of forward / backward compatibility both solutions are equivalent. Only a new parameter is required within the KeyArgs passed to create and commit Key.

If an upgraded server is rolled back, it will still be able to deal with an openKey entry containing expectedGeneration, but it will not process it atomically.

Scope

The intention is to first implement this for OBS buckets. Then address FSO buckets. FSO bucket handling will reuse the same fields, but the handlers on OM are different. We also need to decide on what should happen if a key is renamed or moved folders during the rewrite.

Multi-part keys need more investigation and hence are also excluded in the initial version.

Changes Required

In order to enable the above steps on Ozone, several small changes are needed.

Wire Protocol

  1. The expectedGeneration needs to be added to the KeyInfo protobuf object so it can be stored in the openKey table.
  2. The expectedGeneration needs to be added to the keyArgs protobuf object, which is passed from the client to OM when creating a key.

No new messages need to be defined.

On OM

No new OM handlers are needed. The existing OpenKey and CommitKey handlers will receive the new expectedGeneration and perform the checks.

No new locks are needed on OM. As part of the openKey and commitKey, there are existing locks taken to ensure the key open / commit is atomic. The new checks are performed under those locks, and come down to a couple of long comparisons, so add negligible overhead.

On The Client

  1. We need to allow the updateID (called generation on the client) of an existing key to be accessible when an existing details are read, by adding it to OzoneKey and OzoneKeyDetails. There are internal object changes and do no impact any APIs.
  2. To pass the expectedGeneration to OM on key open, it would be possible to overload the existing OzoneBucket.createKey() method, which already has several overloaded versions, or create a new explicit method on Ozone bucket called rewriteKey, passing the expectedGeneration, eg:
public OzoneOutputStream rewriteKey(String volumeName, String bucketName, String keyName, long size, long expectedGeneration, ReplicationConfig replicationConfigOfNewKey)
     throws IOException 
     
// Can also add an overloaded version of these methods to pass a metadata map, as with the existing
// create key method.        

This specification is roughly in line with the exiting createKey method:

  public OzoneOutputStream createKey(
      String volumeName, String bucketName, String keyName, long size,
      ReplicationConfig replicationConfig,
      Map<String, String> metadata)

An alternative, is to create a new overloaded createKey, but it is probably less confusing to have the new rewriteKey method:

  public OzoneOutputStream createKey(
      String volumeName, String bucketName, String keyName, long size,
      ReplicationConfig replicationConfig, long expectedUpdateID)

The intended usage of this API, is that the existing key details are read, perhaps inspected and then used to open the new key, and then data is written. In this example, the key is overwritten with the same data in a different replication format. Equally, the key could be rewritten with the original data modified in some application specific way. The atomic check guarantees against lost updates if another application thread is attempting to update the same key in a different way.

OzoneKeyDetails exisitingKey = bucket.getKey(keyName);
// Insepect the key and decide if overwrite is desired:
boolean shouldOverwrite = ...
if (shouldOverwrite) {
  try (OutputStream os = bucket.rewriteKey(existingKey.getBucket, existingKey.getVolume, 
    existingKey.getKeyName, existingKey.getSize(), existingKey.getGeneration(), newRepConfig) {
  os.write(bucket.readKey(keyName))
}

Upgrade and Compatibility

Client Server Protocol

If a newer client is talking to an older server, it could call the new atomic API but the server will ignore it without error. The client server versioning framework can be used to avoid this problem.

No new protobuf messages are needed and hence no new Client to OM APIs as the existing APIs are used with an additional parameter.

A single extra field is added to the KeyArgs object, which is passed from the client to OM on key open and commit. This is a new field, so it will be null if not set, and the server will ignore it if it does not expect it.

Disk Layout

A single extra field is added to the OMKeyInfo object which is stored in the openKey table. This is a new field, so it will be null if not set, and the server will ignore it if it does not expect it.

There should be no impact on upgrade / downgrade with the new field added in this way.

Other Storage Systems

Amazon S3 does not offer a facility like this.

Google Cloud has a concept of a generationID which is used in various API calls.

Further Ideas

The intention of this initial design is to make as few changes to Ozone as possible to enable overwriting a key if it has not changed.

It would be possible to have separate generation IDs for metadata changes and data changes to give a more fine grained approach.

It would also be possible to expose these IDs over the S3 interface as well as the Java interface.

However both these options required more changes to Ozone and more API surface to test and support.

The changes suggested here are small, and carry little risk to existing operations if the new field is not passed. They also do not rule out extending the idea to cover a separate metadata generation if such a thing is desired by enough users.