diff options
Diffstat (limited to 'doc/admin-guide/en-US/markdown/glossary.md')
| -rw-r--r-- | doc/admin-guide/en-US/markdown/glossary.md | 300 | 
1 files changed, 0 insertions, 300 deletions
diff --git a/doc/admin-guide/en-US/markdown/glossary.md b/doc/admin-guide/en-US/markdown/glossary.md deleted file mode 100644 index 496d0a428d4..00000000000 --- a/doc/admin-guide/en-US/markdown/glossary.md +++ /dev/null @@ -1,300 +0,0 @@ -Glossary -======== - -**Brick** -:   A Brick is the basic unit of storage in GlusterFS, represented by an export -    directory on a server in the trusted storage pool. -    A brick is expressed by combining a server with an export directory in the following format: - -        `SERVER:EXPORT` -    For example: -        `myhostname:/exports/myexportdir/` - -**Volume** -:   A volume is a logical collection of bricks. Most of the gluster -    management operations happen on the volume. - - -**Subvolume** -:   A brick after being processed by at least one translator or in other words -    set of one or more xlator stacked together is called a sub-volume. - - -**Volfile** -:   Volume (vol) files are configuration files that determine the behavior of the -    GlusterFs trusted storage pool. Volume file is a textual representation of a -    collection of modules (also known as translators) that together implement the -    various functions required. The collection of modules are arranged in a graph-like -    fashion. E.g, A replicated volume's volfile, among other things, would have a -    section describing the replication translator and its tunables. -    This section describes how the volume would replicate data written to it. -    Further, a client process that serves a mount point, would interpret its volfile -    and load the translators described in it. While serving I/O, it would pass the -    request to the collection of modules in the order specified in the volfile. - -    At a high level,  GlusterFs has three entities,that is, Server, Client and Management daemon. -    Each of these entities have their own volume files. -    Volume files for servers and clients are generated by the management daemon -    after the volume is created. - -    Server and Client Vol files are located in /var/lib/glusterd/vols/VOLNAME directory. -    The management daemon vol file is named as glusterd.vol and is located in /etc/glusterfs/ -    directory. - -**glusterd** -:   The daemon/service that manages volumes and cluster membership. It is required to -    run on all the servers in the trusted storage pool. - -**Cluster** -:   A trusted pool of linked computers working together, resembling a single computing resource. -    In GlusterFs, a cluster is also referred to as a trusted storage pool. - -**Client** -:   Any machine that mounts a GlusterFS volume. Any applications that use libgfapi access -    mechanism can also be treated as clients in GlusterFS context. - - -**Server** -:   The machine (virtual or bare metal) that hosts the bricks in which data is stored. - - -**Block Storage** -:   Block special files, or block devices, correspond to devices through which the system moves -    data in the form of blocks. These device nodes often represent addressable devices such as -    hard disks, CD-ROM drives, or memory regions. GlusterFS requires a filesystem (like XFS) that -    supports extended attributes. - - - -**Filesystem** -:   A method of storing and organizing computer files and their data. -    Essentially, it organizes these files into a database for the -    storage, organization, manipulation, and retrieval by the computer's -    operating system. - -    Source: [Wikipedia][] - -**Distributed File System** -:   A file system that allows multiple clients to concurrently access data which is spread across -    servers/bricks in a trusted storage pool. Data sharing among multiple locations is fundamental -    to all distributed file systems. - -**Virtual File System (VFS) -    VFS is a kernel software layer which handles all system calls related to the standard Linux file system. -    It provides a common interface to several kinds of file systems. - -**POSIX** -:   Portable Operating System Interface (for Unix) is the name of a -    family of related standards specified by the IEEE to define the -    application programming interface (API), along with shell and -    utilities interfaces for software compatible with variants of the -    Unix operating system. Gluster exports a fully POSIX compliant file -    system. - -**Extended Attributes** -:   Extended file attributes (abbreviated xattr) is a filesystem feature -    that enables users/programs to associate files/dirs with metadata. - - -**FUSE** -:   Filesystem in Userspace (FUSE) is a loadable kernel module for -    Unix-like computer operating systems that lets non-privileged users -    create their own filesystems without editing kernel code. This is -    achieved by running filesystem code in user space while the FUSE -    module provides only a "bridge" to the actual kernel interfaces. - -    Source: [Wikipedia][1] - - -**GFID** -:   Each file/directory on a GlusterFS volume has a unique 128-bit number -    associated with it called the GFID. This is analogous to inode in a -    regular filesystem. - - -**Infiniband** -    InfiniBand is a switched fabric computer network communications link -    used in high-performance computing and enterprise data centers. - -**Metadata** -:   Metadata is data providing information about one or more other -    pieces of data. - -**Namespace** -:   Namespace is an abstract container or environment created to hold a -    logical grouping of unique identifiers or symbols. Each Gluster -    volume exposes a single namespace as a POSIX mount point that -    contains every file in the cluster. - -**Node** -:   A server or computer that hosts one or more bricks. - -**Open Source** -:   Open source describes practices in production and development that -    promote access to the end product's source materials. Some consider -    open source a philosophy, others consider it a pragmatic -    methodology. - -    Before the term open source became widely adopted, developers and -    producers used a variety of phrases to describe the concept; open -    source gained hold with the rise of the Internet, and the attendant -    need for massive retooling of the computing source code. - -    Opening the source code enabled a self-enhancing diversity of -    production models, communication paths, and interactive communities. -    Subsequently, a new, three-word phrase "open source software" was -    born to describe the environment that the new copyright, licensing, -    domain, and consumer issues created. - -    Source: [Wikipedia][2] - -**Petabyte** -:   A petabyte (derived from the SI prefix peta- ) is a unit of -    information equal to one quadrillion (short scale) bytes, or 1000 -    terabytes. The unit symbol for the petabyte is PB. The prefix peta- -    (P) indicates a power of 1000: - -    1 PB = 1,000,000,000,000,000 B = 10005 B = 1015 B. - -    The term "pebibyte" (PiB), using a binary prefix, is used for the -    corresponding power of 1024. - -    Source: [Wikipedia][3] - - - -**Quorum** -:   The configuration of quorum in a trusted storage pool determines the -    number of server failures that the trusted storage pool can sustain. -    If an additional failure occurs, the trusted storage pool becomes -    unavailable. - -**Quota** -:   Quota allows you to set limits on usage of disk space by directories or -    by volumes. - -**RAID** -:   Redundant Array of Inexpensive Disks (RAID) is a technology that -    provides increased storage reliability through redundancy, combining -    multiple low-cost, less-reliable disk drives components into a -    logical unit where all drives in the array are interdependent. - -**RDMA** -:   Remote direct memory access (RDMA) is a direct memory access from the -    memory of one computer into that of another without involving either -    one's operating system. This permits high-throughput, low-latency -    networking, which is especially useful in massively parallel computer -    clusters. - -**Rebalance** -:   A process of fixing layout and resdistributing data in a volume when a -    brick is added or removed. - -**RRDNS** -:   Round Robin Domain Name Service (RRDNS) is a method to distribute -    load across application servers. RRDNS is implemented by creating -    multiple A records with the same name and different IP addresses in -    the zone file of a DNS server. - -**Samba** -:   Samba allows file and print sharing between computers running Windows and -    computers running Linux. It is an implementation of several services and -    protocols including SMB and CIFS. - -**Self-Heal** -:   The self-heal daemon that runs in the background, identifies -    inconsistencies in files/dirs in a replicated volume and then resolves -    or heals them. This healing process is usually required when one or more -    bricks of a volume goes down and then comes up later. - -**Split-brain** -:   This is a situation where data on two or more bricks in a replicated -    volume start to diverge in terms of content or metadata. In this state, -    one cannot determine programitically which set of data is "right" and -    which is "wrong". - -**Translator** -:   Translators (also called xlators) are stackable modules where each -    module has a very specific purpose. Translators are stacked in a -    hierarchical structure called as graph. A translator receives data -    from its parent translator, performs necessary operations and then -    passes the data down to its child translator in hierarchy. - -**Trusted Storage Pool** -:   A storage pool is a trusted network of storage servers. When you -    start the first server, the storage pool consists of that server -    alone. - -**Scale-Up Storage** -:   Increases the capacity of the storage device in a single dimension. -    For example, adding additional disk capacity to an existing trusted storage pool. - -**Scale-Out Storage** -    Scale out systems are designed to scale on both capacity and performance. -    It increases the capability of a storage device in single dimension. -    For example, adding more systems of the same size, or adding servers to a trusted storage pool -    that increases CPU, disk capacity, and throughput for the trusted storage pool. - -**Userspace** -:   Applications running in user space don’t directly interact with -    hardware, instead using the kernel to moderate access. Userspace -    applications are generally more portable than applications in kernel -    space. Gluster is a user space application. - - -**Geo-Replication** -:   Geo-replication provides a continuous, asynchronous, and incremental -    replication service from site to another over Local Area Networks -    (LAN), Wide Area Network (WAN), and across the Internet. - -**N-way Replication** -:   Local synchronous data replication which is typically deployed across campus -    or Amazon Web Services Availability Zones. - -**Distributed Hash Table Terminology** -**Hashed subvolume** -:   A Distributed Hash Table Translator subvolume to which the file or directory name is hashed to. - -**Cached subvolume** -:   A Distributed Hash Table Translator subvolume where the file content is actually present. -    For directories, the concept of cached-subvolume is not relevant. It is loosely used to mean -    subvolumes which are not hashed-subvolume. - -**Linkto-file** - -:   For a newly created file, the hashed and cached subvolumes are the same. -    When directory entry operations like rename (which can change the name and hence hashed -    subvolume of the file) are performed on the file, instead of moving the entire data in the file -    to a new hashed subvolume, a file is created with the same name on the newly hashed subvolume. -    The purpose of this file is only to act as a pointer to the node where the data is present. -    In the extended attributes of this file, the name of the cached subvolume is stored. -    This file on the newly hashed-subvolume is called a linkto-file. -    The linkto file is relevant only for non-directory entities. - -**Directory Layout** -:   The directory layout specifies the hash-ranges of the subdirectories of a directory to which -    subvolumes they correspond to. - -**Properties of directory layouts:** -:   The layouts are created at the time of directory creation and are persisted as extended attributes -    of the directory. -    A subvolume is not included in the layout if it remained offline at the time of directory creation -    and no directory entries ( such as files and directories) of that directory are created on -    that subvolume. The subvolume is not part of the layout until the fix-layout is complete -    as part of running the rebalance command. If a subvolume is down during access (after directory creation), -    access to any files that hash to that subvolume fails. - -**Fix Layout** -:   A command that is executed during the rebalance process. -    The rebalance process itself comprises of two stages: -    Fixes the layouts of directories to accommodate any subvolumes that are added or removed. -    It also heals the directories, checks whether the layout is non-contiguous, and persists the -    layout in extended attributes, if needed. It also ensures that the directories have the same -    attributes across all the subvolumes. - -    Migrates the data from the cached-subvolume to the hashed-subvolume. - -  [Wikipedia]: http://en.wikipedia.org/wiki/Filesystem -  [1]: http://en.wikipedia.org/wiki/Filesystem_in_Userspace -  [2]: http://en.wikipedia.org/wiki/Open_source -  [3]: http://en.wikipedia.org/wiki/Petabyte  | 
