diff options
author | Joseph Fernandes <josferna@redhat.com> | 2015-05-08 15:18:06 +0530 |
---|---|---|
committer | Humble Devassy Chirammal <humble.devassy@gmail.com> | 2015-05-08 04:33:49 -0700 |
commit | 9c21c5a632ba22a6f46d037bd4fa4d825b24d07f (patch) | |
tree | 304bdc41b3a46388017bb0afccb999d713100575 /doc/features/tier/tier.md | |
parent | a5d124728640f58fa3f90007c149ad69809b6ca4 (diff) |
tiering: Correction to tiering documentation
1) convert to md format
2) Add info about ctr and libgfdb
Change-Id: I531d8a0bff8195f759302c5e613c7af2113729eb
BUG: 1218638
Signed-off-by: Joseph Fernandes <josferna@redhat.com>
Reviewed-on: http://review.gluster.org/10665
Reviewed-by: Humble Devassy Chirammal <humble.devassy@gmail.com>
Tested-by: Humble Devassy Chirammal <humble.devassy@gmail.com>
Diffstat (limited to 'doc/features/tier/tier.md')
-rw-r--r-- | doc/features/tier/tier.md | 168 |
1 files changed, 168 insertions, 0 deletions
diff --git a/doc/features/tier/tier.md b/doc/features/tier/tier.md new file mode 100644 index 00000000000..13e7d971bdf --- /dev/null +++ b/doc/features/tier/tier.md @@ -0,0 +1,168 @@ +##Tiering + +* ####Feature page: +http://www.gluster.org/community/documentation/index.php/Features/data-classification + +* #####Design: goo.gl/bkU5qv + +###Theory of operation + + +The tiering feature enables different storage types to be used by the same +logical volume. In Gluster 3.7, the two types are classified as "cold" and +"hot", and are represented as two groups of bricks. The hot group acts as +a cache for the cold group. The bricks within the two groups themselves are +arranged according to standard Gluster volume conventions, e.g. replicated, +distributed replicated, or dispersed. + +A normal gluster volume can become a tiered volume by "attaching" bricks +to it. The attached bricks become the "hot" group. The bricks within the +original gluster volume are the "cold" bricks. + +For example, the original volume may be dispersed on HDD, and the "hot" +tier could be distributed-replicated SSDs. + +Once this new "tiered" volume is built, I/Os to it are subjected to cacheing +heuristics: + +* All I/Os are forwarded to the hot tier. + +* If a lookup fails to the hot tier, the I/O will be forwarded to the cold +tier. This is a "cache miss". + +* Files on the hot tier that are not touched within some time are demoted +(moved) to the cold tier (see performance parameters, below). + +* Files on the cold tier that are touched one or more times are promoted +(moved) to the hot tier. (see performance parameters, below). + +This resembles implementations by Ceph and the Linux data management (DM) +component. + +Performance enhancements being considered include: + +* Biasing migration of large files over small. + +* Only demoting when the hot tier is close to full. + +* Write-back cache for database updates. + +###Code organization + +The design endevors to be upward compatible with future migration policies, +such as scheduled file migration, data classification, etc. For example, +the caching logic is self-contained and separate from the file migration. A +different set of migration policies could use the same underlying migration +engine. The I/O tracking and meta data store compontents are intended to be +reusable for things besides caching semantics. + +####Libgfdb: + +Libgfdb provides abstract mechanism to record extra/rich metadata +required for data maintenance, such as data tiering/classification. +It provides consumer with API for recording and querying, keeping +the consumer abstracted from the data store used beneath for storing data. +It works in a plug-and-play model, where data stores can be plugged-in. +Presently we have plugin for Sqlite3. In the future will provide recording +and querying performance optimizer. In the current implementation the schema +of metadata is fixed. + +####Schema: + + GF_FILE_TB Table: + This table has one entry per file inode. It holds the metadata required to + make decisions in data maintenance. + GF_ID (Primary key) : File GFID (Universal Unique IDentifier in the namespace) + W_SEC, W_MSEC : Write wind time in sec & micro-sec + UW_SEC, UW_MSEC : Write un-wind time in sec & micro-sec + W_READ_SEC, W_READ_MSEC : Read wind time in sec & micro-sec + UW_READ_SEC, UW_READ_MSEC : Read un-wind time in sec & micro-sec + WRITE_FREQ_CNTR INTEGER : Write Frequency Counter + READ_FREQ_CNTR INTEGER : Read Frequency Counter + + GF_FLINK_TABLE: + This table has all the hardlinks to a file inode. + GF_ID : File GFID (Composite Primary Key)``| + GF_PID : Parent Directory GFID (Composite Primary Key) |-> Primary Key + FNAME : File Base Name (Composite Primary Key)__| + FPATH : File Full Path (Its redundant for now, this will go) + W_DEL_FLAG : This Flag is used for crash consistancy, when a link is unlinked. + i.e Set to 1 during unlink wind and during unwind this record is deleted + LINK_UPDATE : This Flag is used when a link is changed i.e rename. + Set to 1 when rename wind and set to 0 in rename unwind + +Libgfdb API : +Refer libglusterfs/src/gfdb/gfdb_data_store.h + +####ChangeTimeRecorder (CTR) Translator: + +ChangeTimeRecorder(CTR) is server side xlator(translator) which sits +just above posix xlator. The main role of this xlator is to record the +access/write patterns on a file residing the brick. It records the +read(only data) and write(data and metadata) times and also count on +how many times a file is read or written. This xlator also captures +the hard links to a file(as its required by data tiering to move +files). + +CTR Xlator is the consumer of libgfdb. + +To Enable/Disable CTR Xlator: + + **gluster volume set <volume-name> features.ctr-enabled {on/off}** + +To Enable/Disable Frequency Counter Recording in CTR Xlator: + + **gluster volume set <volume-name> features.record-counters {on/off}** + + +####Migration daemon: + +When a tiered volume is created, a migration daemon starts. There is one daemon +for every tiered volume per node. The daemon sleeps and then periodically +queries the database for files to promote or demote. The query callbacks +assembles files in a list, which is then enumerated. The frequencies by +which promotes and demotes happen is subject to user configuration. + +Selected files are migrated between the tiers using existing DHT migration +logic. The tier translator will leverage DHT rebalance performance +enhancements. + +Configurable for Migration daemon: + + gluster volume set <volume-name> cluster.tier-demote-frequency <SECS> + + gluster volume set <volume-name> cluster.tier-promote-frequency <SECS> + + gluster volume set <volume-name> cluster.read-freq-threshold <SECS> + + gluster volume set <volume-name> cluster.write-freq-threshold <SECS> + + +####Tier Translator: + +The tier translator is the root node in tiered volumes. The first subvolume +is the cold tier, and the second the hot tier. DHT logic for fowarding I/Os +is largely unchanged. Exceptions are handled according to the dht_methods_t +structure, which forks control according to DHT or tier type. + +The major exception is DHT's layout is not utilized for choosing hashed +subvolumes. Rather, the hot tier is always the hashed subvolume. + +Changes to DHT were made to allow "stacking", i.e. DHT over DHT: + +* readdir operations remember the index of the "leaf node" in the volume graph +(client id), rather than a unique index for each DHT instance. + +* Each DHT instance uses a unique extended attribute for tracking migration. + +* In certain cases, it is legal for tiered volumes to have unpopulated inodes +(wheras this would be an error in DHT's case). + +Currently tiered volume expansion (adding and removing bricks) is unsupported. + +####glusterd: + +The tiered volume tree is a composition of two other volumes. The glusterd +daemon builds it. Existing logic for adding and removing bricks is heavily +leveraged to attach and detach tiers, and perform statistics collection. + |