You are viewing the documentation of an older version of the TiDB database (TiDB v2.1).

It is recommended that you use the latest LTS version of the TiDB database.

Software and Hardware Recommendations


As an open source distributed NewSQL database with high performance, TiDB can be deployed in the Intel architecture server and major virtualization environments and runs well. TiDB supports most of the major hardware networks and Linux operating systems.

Linux OS version requirements

Linux OS PlatformVersion
Red Hat Enterprise Linux7.3 or later
CentOS7.3 or later
Oracle Enterprise Linux7.3 or later
Ubuntu LTS16.04 or later
  • For Oracle Enterprise Linux, TiDB supports the Red Hat Compatible Kernel (RHCK) and does not support the Unbreakable Enterprise Kernel provided by Oracle Enterprise Linux.
  • A large number of TiDB tests have been run on the CentOS 7.3 system, and in our community there are a lot of best practices in which TiDB is deployed on the Linux operating system. Therefore, it is recommended to deploy TiDB on CentOS 7.3 or later.
  • The support for the Linux operating systems above includes the deployment and operation in physical servers as well as in major virtualized environments like VMware, KVM and XEN.

Server recommendations

You can deploy and run TiDB on the 64-bit generic hardware server platform in the Intel x86-64 architecture. The requirements and recommendations about server hardware configuration (ignoring the resources occupied by the operating system itself) for development, test, and production environments are as follows:

Development and test environments

ComponentCPUMemoryLocal StorageNetworkInstance Number (Minimum Requirement)
TiDB8 core+16 GB+No special requirementsGigabit network card1 (can be deployed on the same machine with PD)
PD4 core+8 GB+SAS, 200 GB+Gigabit network card1 (can be deployed on the same machine with TiDB)
TiKV8 core+32 GB+SAS, 200 GB+Gigabit network card3
Total Server Number4
  • In the test environment, the TiDB and PD instances can be deployed on the same server.
  • For performance-related test, do not use low-performance storage and network hardware configuration, in order to guarantee the correctness of the test result.
  • For the TiKV server, it is recommended to use NVMe SSDs to ensure faster reads and writes.
  • The TiDB server uses the disk to store server logs, so there are no special requirements for the disk type and capacity in the test environment.

Production environment

ComponentCPUMemoryHard Disk TypeNetworkInstance Number (Minimum Requirement)
TiDB16 core+32 GB+SAS10 Gigabit network card (2 preferred)2
PD4 core+8 GB+SSD10 Gigabit network card (2 preferred)3
TiKV16 core+32 GB+SSD10 Gigabit network card (2 preferred)3
Monitor8 core+16 GB+SASGigabit network card1
Total Server Number9
  • In the production environment, the TiDB and PD instances can be deployed on the same server. If you have a higher requirement for performance and reliability, try to deploy them separately.
  • It is strongly recommended to use higher configuration in the production environment.
  • It is recommended to keep the size of TiKV hard disk within 2 TB if you are using PCIe SSDs or within 1.5 TB if you are using regular SSDs.

Network requirements

As an open source distributed NewSQL database, TiDB requires the following network port configuration to run. Based on the TiDB deployment in actual environments, the administrator can open relevant ports in the network side and host side.

ComponentDefault PortDescription
TiDB4000the communication port for the application and DBA tools
TiDB10080the communication port to report TiDB status
TiKV20160the TiKV communication port
TiKV20180the communication port to report TiKV status
PD2379the communication port between TiDB and PD
PD2380the inter-node communication port within the PD cluster
Pump8250the Pump communication port
Drainer8249the Drainer communication port
Prometheus9090the communication port for the Prometheus service
Pushgateway9091the aggregation and report port for TiDB, TiKV, and PD monitor
Node_exporter9100the communication port to report the system information of every TiDB cluster node
Blackbox_exporter9115the Blackbox_exporter communication port, used to monitor the ports in the TiDB cluster
Grafana3000the port for the external Web monitoring service and client (Browser) access
Grafana8686the grafana_collector communication port, used to export the Dashboard as the PDF format
Kafka_exporter9308the Kafka_exporter communication port, used to monitor the binlog Kafka cluster

Web browser requirements

TiDB relies on Grafana to provide visualization of database metrics. A recent version of Internet Explorer, Chrome or Firefox with Javascript enabled is sufficient.

Was this page helpful?