-
Notifications
You must be signed in to change notification settings - Fork 0
PDP 28: Cross routing key ordering
Today on the reader, when data is coming from multiple segments, there is no way to maintain even approximate order. Events are pulled from segment and given to the reading application based what segment has more data in memory. This can result in some parts of the keyspace getting very far ahead of others. Unless they merge down to a common segment there will never be any event that causes them to adjust to be closer to one another. So the longer a reader group has been running the more out of order its events become. While we do not provide any guarantees that this violates, it represents a significant inconvenience to many applications.
To achieve best effort ordering on a cross routing key basis.
This proposal is narrow in focus and is not attempting to solve some related problems:
- Put hard bounds on how out of order events can be.
- Provide a solution for event time.
- Provide a solution for watermarking in the Flink connector.
These are important in their own right, and may be relevant to the discussion of adoption, but are out of scope for the proposal itself.
A time index is stored on the server, and the values passed to the client. These are used to determine the order of events on a cross segment basis and provide weak fairness.
None.
In the SegmentRead reply from the server to the client we would add an additional field that would contain a number indicating how the data returned should be ordered relative to other data. It would additionally have a token for where the read left off.
In StreamSegmentInfo
we add a field that contains the current value that is to be used for data if it were to be added to the segment now.
In ReadSegment
we would add an additional field of the token that was returned from the previous SegmentRead
.
This change would be backwards compatible, as we can just infer a value of 0 for the ordering field if the field is missing. This would prevent the feature from working, but would be no different than if it did not exist. If the token is missing this means the server does not have the new feature and the client will simply not provide the token it it's ReadSegment call.
None.
The reader keeps a buffer from each segment that it is reading from and selects which one to read from in response to a readNextEvent()
call using a component called Orderer. Currently the only criteria this class takes into account is the amount of data in memory.
In SegmentInputStreamImpl
on the client, it would track the ordering information returned from the server and expose the current value via its interface so that Orderer could check if one segment is ahead or behind another. A reasonable initial implementation would be for the orderer to select the segment returning the lowest value.
Additionally, in AsyncSegmentInputStreamImpl
, it would hold in memory the token it received from the last SegmentRead and pass it into the next ReadSegment request.
In EventStreamReaderImpl
, the getLag()
function can return the data from the time index. This will fulfill Issue 191. The getLag()
function is already used by the ReaderGroupStateManager
to determine how to rebalance segments, but this would anchor it to be much more accurate.
In every StreamSegmentAppendOperation
(on non-transaction segments) and every MergeTransactionOperation
written to Tier-1 we would add an additional long. This value would be obtained from a monotonic clock class.
The clock would be shared across segment and be at approximately the same value across hosts. This could be derived from a wall clock enforcing monotonically, or be pulled from a central source and tracked using nanotime for deltas.
In OperationProcessor
, we would update a new component of metadata for the segment on append. This would consist of offset, clock pairs using the value that was written to tier-1 for the last DataFrame
of the append. The ReadIndex
can store this in RocksDB as it is non-constant in size. We can use the concatenate operator to keep the stored content concise.
This data constitutes an event offset to time index. The ReadIndex
can use this to locate the time for an event being read. When a ReadSegment
request arrives with a token, that token can be interpreted as a offset into this index data. Then the entries for the relevant range for the index can be returned along with a token for where the next read should occur at.
When StorageWriter
moves data to Tier-2 the corresponding time index should also be moved there and the ReadIndex
updated. The data can be written as its own file. Once the data is written to tier-2 it can be evicted from memory just like the data in the segment.
If a read arrives at PravegaRequestProcessor
from a new client, it will not have a token. If the offset is 0 the the token can be assumed to be zero. However, if it is not, we can locate the correct offset with a binary or better yet a Newtonian search. (If the segment is 10MB in length and we are being asked to read at 8.5MB in, instead of selecting a guess 50% of the way between the start and end of the index, pick one 85% of the way.)
Locating an initial token does not require introducing a new API on the store, we can simply have a token constant that represents that the client does not have a token and the value needs to be located.
We need not track index data on transaction segments. This is because if we were to track it there and merge it, it would violate the monotonically of the values. Instead we should use a single point for the transaction at the time of merge. So the MergeTransactionOperation would need to contain the a value from the clock.
Pravega - Streaming as a new software defined storage primitive
- Contributing
- Guidelines for committers
- Testing
-
Pravega Design Documents (PDPs)
- PDP-19: Retention
- PDP-20: Txn timeouts
- PDP-21: Protocol revisioning
- PDP-22: Bookkeeper based Tier-2
- PDP-23: Pravega Security
- PDP-24: Rolling transactions
- PDP-25: Read-Only Segment Store
- PDP-26: Ingestion Watermarks
- PDP-27: Admin Tools
- PDP-28: Cross routing key ordering
- PDP-29: Tables
- PDP-30: Byte Stream API
- PDP-31: End-to-end Request Tags
- PDP-32: Controller Metadata Scalability
- PDP-33: Watermarking
- PDP-34: Simplified-Tier-2
- PDP-35: Move controller metadata to KVS
- PDP-36: Connection pooling
- PDP-37: Server-side compression
- PDP-38: Schema Registry
- PDP-39: Key-Value Tables
- PDP-40: Consistent order guarantees for storage flushes
- PDP-41: Enabling Transport Layer Security (TLS) for External Clients
- PDP-42: New Resource String Format for Authorization
- PDP-43: Large Events
- PDP-44: Lightweight Transactions
- PDP-45: Healthcheck
- PDP-46: Read Only Permissions For Reading Data
- PDP-47: Pravega Message Queues