Keep storage tiers - OBSOLETE¶
WARNING For historical reasons only, superseded by Keep Storage Classes and never implemented WARNING
Typically, an Arvados cluster has access to multiple storage devices with different cost/performance trade-offs.
- Local SSD
- Local HDD
- Object storage service provided by cloud vendor
- Slower or less reliable object storage service provided by same cloud vendor
Users should be able to specify a minimum storage tier for each collection. Arvados should ensure that every data block referenced by a collection is stored at the specified tier or better. The cluster administrator should be able to specify a default tier, and assign a tier number to each storage device. It should be possible to configure multiple storage devices at the same tier: for example, this allows blocks to be distributed more or less uniformly across several (equivalent) cloud storage buckets for performance reasons.
Storage tier features (and implementation) are similar to replication-level features.
Each Keep volume has an integer parameter, "tier". Interpretation is site-specific, except that when M≤N, tier M can satisfy a requirement for tier N, i.e., smaller tier numbers are better. Some volume drivers are capable of discovering the tier number for a volume by inspecting the underlying storage device (e.g., a cloud storage bucket) but in all cases a sysadmin can specify a value.
There is a site-wide default tier number which is used for collections that do not specify a desired tier. Typically this is tier 1.
Storing data at a non-default tier¶
Tools that write data to Keep should allow the caller to specify a storage tier. The desired tier is sent to Keep services as a header (X-Keep-Desired-Tier) with each write request. Keep services return an error when the data cannot be written to the requested tier (or better).
Moving data between tiers¶
Each collection has an integer field, "tier_desired". If tier_desired is not null, all blocks referenced by the collection should be stored at the given tier (or better).
Keep-balance tracks the maximum allowed tier for each block, and moves blocks between tiers as needed. The strategy is similar to fixing rendezvous probe order: if a block is stored at the wrong tier, a new copy is made at the correct tier; then, in a subsequent balancing operation, the redundant copy is detected and deleted. This increases the danger of data loss due to races between concurrent keep-balance processes. Keep-balance should have a reliable way to detect/avoid concurrent balancing operations.
(Note: the following section uses the term "mount" to mean what the keepstore code base calls a "volume": i.e., an attachment of a storage device to a keepstore process.)Keepstore provides APIs that let keep-balance see which blocks are stored on which mount points, and copy/delete blocks to/from specific mount points:
- Get current mounts
- List blocks on a specified mount
- Delete block from a specified mount
- Pull block to a specified mount
Reporting¶After each rebalance operation, keep-balance logs a summary of discrepancies between actual and desired allocation of blocks to storage tiers. Examples:
- N blocks (M bytes) are stored at tier 3 but are referenced by collections at tier 2.
- N blocks (M bytes) are stored at tier 1 but are not referenced by any collections at tier T<2.