Glossary
This glossary provides definitions for key terms related to the TiDB platform.
Other available glossaries:
A
ACID
ACID refers to the four key properties of a transaction: atomicity, consistency, isolation, and durability. Each of these properties is described below.
Atomicity means that either all the changes of an operation are performed, or none of them are. TiDB ensures the atomicity of the Region that stores the Primary Key to achieve the atomicity of transactions.
Consistency means that transactions always bring the database from one consistent state to another. In TiDB, data consistency is ensured before writing data to the memory.
Isolation means that a transaction in process is invisible to other transactions until it completes. This allows concurrent transactions to read and write data without sacrificing consistency. For more information, see TiDB transaction isolation levels.
Durability means that once a transaction is committed, it remains committed even in the event of a system failure. TiKV uses persistent storage to ensure durability.
B
Backup & Restore (BR)
BR is the backup and restore tool for TiDB. For more information, see BR Overview.
br
is the br command line tool used for backups or restores in TiDB.
Baseline Capturing
Baseline Capturing captures queries that meet capturing conditions and create bindings for them. It is used for preventing regression of execution plans during an upgrade.
Batch Create Table
The Batch Create Table feature greatly speeds up the creation of multiple tables at a time by creating tables in batches. For example, when restoring thousands of tables using the Backup & Restore (BR) tool, this feature helps reduce the overall recovery time. For more information, see Batch Create Table.
Bucket
A Region is logically divided into several small ranges called bucket. TiKV collects query statistics by buckets and reports the bucket status to PD. For more information, see the Bucket design doc.
C
Cached Table
With the cached table feature, TiDB loads the data of an entire table into the memory of the TiDB server, and TiDB directly gets the table data from the memory without accessing TiKV, which improves the read performance.
Cluster
A cluster is a group of nodes that work together to provide services. By using clusters in a distributed system, TiDB achieves higher availability and greater scalability compared to a single-node setup.
In the distributed architecture of the TiDB database:
- TiDB nodes provide a scalable SQL layer for client interactions.
- PD nodes provide a resilient metadata layer for TiDB.
- TiKV nodes, using the Raft protocol, provide highly available, scalable, and resilient storage for TiDB.
For more information, see TiDB Architecture.
Coalesce Partition
Coalesce Partition is a way of decreasing the number of partitions in a Hash or Key partitioned table. For more information, see Manage Hash and Key partitions.
Column Family (CF)
In RocksDB and TiKV, a Column Family (CF) represents a logical grouping of key-value pairs within a database.
Common Table Expression (CTE)
A Common Table Expression (CTE) enables you to define a temporary result set that can be referred to multiple times within a SQL statement using the WITH
clause, which improves the statement readability and execution efficiency. For more information, see Common Table Expression.
Continuous Profiling
Continuous Profiling is a way to observe resource overhead at the system call level. With Continuous Profiling, TiDB provides fine-grained observations of performance issues, helping operations teams identify the root cause using a flame graph. For more information, see TiDB Dashboard Instance Profiling - Continuous Profiling.
Coprocessor
Coprocessor is a coprocessing mechanism that shares the computation workload with TiDB. It is located in the storage layer (TiKV or TiFlash) and collaboratively processes computations pushed down from TiDB on a per-Region basis.
D
Dumpling
Dumpling is a data export tool for exporting data stored in TiDB, MySQL, or MariaDB as SQL or CSV data files. It can also be used for logical full backups or exports. Additionally, Dumpling supports exporting data to Amazon S3.
For more information, see Use Dumpling to Export Data.
Data Definition Language (DDL)
Data Definition Language (DDL) is a part of the SQL standard that deals with creating, modifying, and dropping tables and other objects. For more information, see DDL Introduction.
Data Migration (DM)
Data Migration (DM) is a tool for migrating data from MySQL-compatible databases into TiDB. DM reads data from a MySQL-compatible database instance and applies it to a TiDB target instance. For more information, see DM Overview.
Data Modification Language (DML)
Data Modification Language (DML) is a part of the SQL standard that deals with inserting, updating, and dropping rows in tables.
Development Milestone Release (DMR)
Development Milestone Releases (DMR) are TiDB releases that introduce the latest features but do not offer long-term support. For more information, see TiDB Versioning.
Disaster Recovery (DR)
Disaster Recovery (DR) includes solutions that can be used to recover data and services from a disaster in the future. TiDB offers various Disaster Recovery solutions such as backups and replication to standby clusters. For more information, see Overview of TiDB Disaster Recovery Solutions.
Distributed eXecution Framework (DXF)
Distributed eXecution Framework (DXF) is the framework used by TiDB to centrally schedule certain tasks (such as creating indexes or importing data) and execute them in a distributed manner. DXF is designed to efficiently use cluster resources while controlling resource usage and reducing the impact on core business transactions. For more information, see DXF Introduction.
Dynamic Pruning
Dynamic pruning mode is one of the modes that TiDB accesses partitioned tables. In dynamic pruning mode, each operator supports direct access to multiple partitions. Therefore, TiDB no longer uses Union. Omitting the Union operation can improve the execution efficiency and avoid the problem of Union concurrent execution.
E
Expression index
The expression index is a special type of index created on an expression. Once an expression index is created, TiDB can use this index for expression-based queries, significantly improving query performance.
For more information, see CREATE INDEX - Expression index.
G
Garbage Collection (GC)
Garbage Collection (GC) is a process that clears obsolete data to free up resources. For information on TiKV GC process, see GC Overview.
General Availability (GA)
General Availability (GA) of a feature means the feature is fully tested and is Generally Available for use in production environments. TiDB features can be released as GA in both DMR and LTS releases. However, as TiDB does not provide patch releases for DMR it is generally recommended to use the LTS release for production use.
Global Transaction Identifiers (GTIDs)
Global Transaction Identifiers (GTIDs) are unique transaction IDs used in MySQL binary logs to track which transactions have been replicated. Data Migration (DM) uses these IDs to ensure consistent replication.
H
Hotspot
Hotspot refers to a situation where the read and write workloads in TiKV are concentrated on one or a few Regions or nodes. This can lead to performance bottlenecks, preventing optimal system performance. To solve hotspot issues, see Troubleshoot Hotspot Issues.
Hybrid Transactional and Analytical Processing (HTAP)
Hybrid Transactional and Analytical Processing (HTAP) is a database feature that enables both OLTP (Online Transactional Processing) and OLAP (Online Analytical Processing) workloads within the same database. For TiDB, the HTAP feature is provided by using TiKV for row storage and TiFlash for columnar storage. For more information, see Quick Start with TiDB HTAP and Explore HTAP.
I
In-Memory Pessimistic Lock
The in-memory pessimistic lock is a new feature introduced in TiDB v6.0.0. When this feature is enabled, pessimistic locks are usually stored in the memory of the Region leader only, and are not persisted to disk or replicated through Raft to other replicas. This feature can greatly reduce the overhead of acquiring pessimistic locks and improve the throughput of pessimistic transactions.
Index Merge
Index Merge is a method introduced in TiDB v4.0 to access tables. Using this method, the TiDB optimizer can use multiple indexes per table and merge the results returned by each index. In some scenarios, this method makes the query more efficient by avoiding full table scans. Since v5.4, Index Merge has become a GA feature.
K
Key Management Service (KMS)
Key Management Service (KMS) enables the storage and retrieval of secret keys in a secure way. Examples include AWS KMS, Google Cloud KMS, and HashiCorp Vault. Various TiDB components can use KMS to manage keys for storage encryption and related services.
Key-Value (KV)
Key-Value (KV) is a way of storing information by associating values with unique keys, allowing quick data retrieval. TiDB uses TiKV to map tables and indexes into key-value pairs, enabling efficient data storage and access across the database.
L
Leader/Follower/Learner
Leader/Follower/Learner each corresponds to a role in a Raft group of peers. The leader services all client requests and replicates data to the followers. If the group leader fails, one of the followers will be elected as the new leader. Learners are non-voting followers that only serves in the process of replica addition.
Lightweight Directory Access Protocol (LDAP)
Lightweight Directory Access Protocol (LDAP) is a standardized way of accessing a directory with information. It is commonly used for account and user data management. TiDB supports LDAP via LDAP authentication plugins.
Lock View
The Lock View feature provides more information about lock conflicts and lock waits in pessimistic locking, making it convenient for DBAs to observe transaction locking situations and troubleshoot deadlock issues.
For more information, see system table documentation: TIDB_TRX
, DATA_LOCK_WAITS
, and DEADLOCKS
.
Long Term Support (LTS)
Long Term Support (LTS) refers to software versions that are extensively tested and maintained for extended periods. For more information, see TiDB Versioning.
M
Massively Parallel Processing (MPP)
Starting from v5.0, TiDB introduces Massively Parallel Processing (MPP) architecture through TiFlash nodes, which shares the execution workloads of large join queries among TiFlash nodes. When the MPP mode is enabled, TiDB, based on cost, determines whether to use the MPP framework to perform the calculation. In the MPP mode, the join keys are redistributed through the Exchange operation while being calculated, which distributes the calculation pressure to each TiFlash node and speeds up the calculation. For more information, see Use TiFlash MPP Mode.
Multi-version concurrency control (MVCC)
MVCC is a concurrency control mechanism in TiDB and other databases. It processes the memory read by transactions to achieve concurrent access to TiDB, thereby avoiding blocking caused by conflicts between concurrent reads and writes.
O
Old value
The "original value" in the incremental change log output by TiCDC. You can specify whether the incremental change log output by TiCDC contains the "original value".
Online Analytical Processing (OLAP)
Online Analytical Processing (OLAP) refers to database workloads focused on analytical tasks, such as data reporting and complex queries. OLAP is characterized by read-heavy queries that process large volumes of data across many rows.
Online Transaction Processing (OLTP)
Online Transaction Processing (OLTP) refers to database workloads focused on transactional tasks, such as selecting, inserting, updating, and deleting small sets of records.
Out of Memory (OOM)
Out of Memory (OOM) is a situation where a system fails due to insufficient memory. For more information, see Troubleshoot TiDB OOM Issues.
Operator
An operator is a collection of actions that applies to a Region for scheduling purposes. Operators perform scheduling tasks such as "migrate the leader of Region 2 to Store 5" and "migrate replicas of Region 2 to Store 1, 4, 5".
An operator can be computed and generated by a scheduler, or created by an external API.
Operator step
An operator step is a step in the execution of an operator. An operator normally contains multiple Operator steps.
Currently, available steps generated by PD include:
TransferLeader
: Transfers leadership to a specified memberAddPeer
: Adds peers to a specified storeRemovePeer
: Removes a peer of a RegionAddLearner
: Adds learners to a specified storePromoteLearner
: Promotes a specified learner to a voting memberSplitRegion
: Splits a specified Region into two
Optimistic transaction
Optimistic transactions are transactions that use optimistic concurrency control and generally do not cause conflicts in concurrent environments. After enabling optimistic transactions, TiDB checks for conflicts only when the transaction is finally committed. The optimistic transaction mode is suitable for read-heavy and write-light concurrent scenarios, which can improve the performance of TiDB.
For more information, see TiDB Optimistic Transaction Model.
P
Partitioning
Partitioning refers to physically dividing a table into smaller table partitions, which can be done by partition methods such as RANGE, LIST, HASH, and KEY partitioning.
PD Control (pd-ctl)
PD Control (pd-ctl) is a command-line tool used to interact with the Placement Driver (PD) in the TiDB cluster. You can use it to obtain cluster status information and modify the cluster configuration. For more information, see PD Control User Guide.
Pending/Down
"Pending" and "down" are two special states of a peer. Pending indicates that the Raft log of followers or learners is vastly different from that of leader. Followers in pending cannot be elected as leader. "Down" refers to a state that a peer ceases to respond to leader for a long time, which usually means the corresponding node is down or isolated from the network.
Placement Driver (PD)
Placement Driver (PD) is a core component in the TiDB Architecture responsible for storing metadata, assigning Timestamp Oracle (TSO) for transaction timestamps, orchestrating data placement on TiKV, and running TiDB Dashboard. For more information, see TiDB Scheduling.
Placement Rules
Placement rules are used to configure the placement of data in a TiKV cluster. With this feature, you can specify the deployment of tables and partitions to different regions, data centers, cabinets, or hosts. Use cases include optimizing data availability strategies at low cost, ensuring that local data replicas are available for local stale reads, and complying with local data compliance requirements.
For more information, see Placement Rules in SQL.
Point Get
Point get means reading a single row of data by a unique index or primary index, the returned resultset is up to one row.
Point in Time Recovery (PITR)
Point in Time Recovery (PITR) enables you to restore data to a specific point in time (for example, just before an unintended DELETE
statement). For more information, see TiDB Log Backup and PITR Architecture.
Predicate columns
In most cases, when executing SQL statements, the optimizer only uses statistics of some columns (such as columns in the WHERE
, JOIN
, ORDER BY
, and GROUP BY
statements). These used columns are called predicate columns. For more information, see Collect statistics on some columns.
Q
Queries Per Second (QPS)
Queries Per Second (QPS) is the number of queries a database service handles per second, serving as a key performance metric for database throughput.
Quota Limiter
Quota Limiter is an experimental feature introduced in TiDB v6.0.0. If the machine on which TiKV is deployed has limited resources, for example, with only 4v CPU and 16 G memory, and the foreground of TiKV processes too many read and write requests, the CPU resources used by the background are occupied to help process such requests, which affects the performance stability of TiKV. To avoid this situation, the quota-related configuration items can be set to limit the CPU resources to be used by the foreground.
R
Raft Engine
Raft Engine is an embedded persistent storage engine with a log-structured design. It is built for TiKV to store multi-Raft logs. Since v5.4, TiDB supports using Raft Engine as the log storage engine. For more information, see Raft Engine.
Region Split
A region in a TiKV cluster is not divided at the beginning but is gradually split as data is written to it. The process is called Region split.
The mechanism of Region split is to use one initial Region to cover the entire key space, and generate new Regions through splitting existing ones every time the size of the Region or the number of keys has reached a threshold.
Region/Peer/Raft Group
Region is the minimal piece of data storage in TiKV, each representing a range of data (256 MiB by default). Each Region has three replicas by default. A replica of a Region is called a peer. Multiple peers of the same Region replicate data via the Raft consensus algorithm, so peers are also members of a Raft instance. TiKV uses Multi-Raft to manage data. That is, for each Region, there is a corresponding, isolated Raft group.
Remote Procedure Call (RPC)
Remote Procedure Call (RPC) is a communication way between software components. In a TiDB cluster, the gRPC standard is used for communication between different components such as TiDB, TiKV, and TiFlash.
Request Unit (RU)
Request Unit (RU) is a unified abstraction unit for the resource usage in TiDB. It is used with Resource Control to manage resource usage.
Restore
Restore is the reverse of the backup operation. It is the process of bringing back the system to an earlier state by retrieving data from a prepared backup.
RocksDB
RocksDB is an LSM-tree structured engine that provides key-value storage and read-write functionality. It was developed by Facebook and is based on LevelDB. RocksDB is the core storage engine of TiKV.
S
Scheduler
Schedulers are components in PD that generate scheduling tasks. Each scheduler in PD runs independently and serves different purposes. The commonly used schedulers are:
balance-leader-scheduler
: Balances the distribution of leadersbalance-region-scheduler
: Balances the distribution of peershot-region-scheduler
: Balances the distribution of hot Regionsevict-leader-{store-id}
: Evicts all leaders of a node (often used for rolling upgrades)
Security Enhanced Mode (SEM)
The Security Enhanced Mode (SEM) is used for finer-grained permission control of TiDB administrators. Inspired by systems such as Security-Enhanced Linux, SEM reduces the abilities of users with the SUPER
privilege and instead requires RESTRICTED
fine-grained privileges, which must be explicitly granted to control specific administrative actions.
For more information, see System Variables documentation - tidb_enable_enhanced_security
.
Stale Read
Stale Read is a mechanism that TiDB applies to read historical versions of data stored in TiDB. Using this mechanism, you can read the corresponding historical data of a specific point in time or within a specified time range, and thus save the latency brought by data replication between storage nodes. When you use Stale Read, TiDB randomly selects a replica for data reading, which means that all replicas are available for data reading.
For more information, see Stale Read.
Static Sorted Table / Sorted String Table (SST)
Static Sorted Table or Sorted String Table is a file storage format used in RocksDB (a storage engine used by TiKV).
Store
A store refers to the storage node in the TiKV cluster (an instance of tikv-server
). Each store has a corresponding TiKV instance.
T
Temporary table
Temporary tables enable you to store intermediate calculation results temporarily, eliminating the need to create and drop tables repeatedly. Once the data is no longer needed, TiDB automatically cleans up and recycles the temporary tables. This feature simplifies application logic, reduces table management overhead, and improves performance.
For more information, see Temporary Tables.
TiCDC
TiCDC is a tool that enables incremental data replication from TiDB to various downstream targets. These downstream targets can include other TiDB instances, MySQL-compatible databases, storage services, and streaming processors (such as Kafka and Pulsar). TiCDC pulls the data change logs from the upstream TiKV, parses them into ordered row-level change data, and then outputs the data to the downstream. For more information about the concepts and terms of TiCDC, see TiCDC Glossary.
TiDB Lightning
TiDB Lightning is a tool for importing Terabyte-level data from static files into TiDB clusters. It is commonly used for the initial data import into TiDB clusters.
For more information on the concepts and terminology of TiDB Lightning, see TiDB Lightning Glossary.
TiFlash
TiFlash is a key component of TiDB's HTAP architecture. It is a columnar extension of TiKV that provides both strong consistency and good isolation. TiFlash maintains columnar replicas by asynchronously replicating data from TiKV using the Raft Learner protocol. For reads, it leverages the Raft consensus index and MVCC (Multi-Version Concurrency Control) to achieve Snapshot Isolation consistency. This architecture effectively addresses isolation and synchronization challenges in HTAP workloads, enabling efficient analytical queries while maintaining real-time data consistency.
Timestamp Oracle (TSO)
Because TiKV is a distributed storage system, it requires a global timing service, Timestamp Oracle (TSO), to assign a monotonically increasing timestamp. In TiKV, such a feature is provided by PD, and in Google Spanner, this feature is provided by multiple atomic clocks and GPS. For more information, see TSO.
TiUP
TiUP is a management tool used for deploying, upgrading, and managing TiDB clusters, as well as managing various components within the TiDB cluster including TiDB, PD, and TiKV. With TiUP, you can easily run any component within TiDB by executing a single command, greatly simplifying the management process.
Top SQL
Top SQL helps locate SQL queries that contribute to a high load of a TiDB or TiKV node in a specified time range. For more information, see Top SQL user document.
Transactions Per Second (TPS)
Transactions Per Second (TPS) is the number of transactions a database processes per second, serving as a key metric for measuring database performance and throughput.
U
Uniform Resource Identifier (URI)
Uniform Resource Identifier (URI) is a standardized format for identifying a resource. For more information, see Uniform Resource Identifier on Wikipedia.
Universally Unique Identifier (UUID)
Universally Unique Identifier (UUID) is a 128-bit (16-byte) generated ID used to uniquely identify records in a database. For more information, see UUID.