Deploy TiDB in General Kubernetes

This document describes how to deploy a TiDB cluster in general Kubernetes.


Deploy the TiDB cluster

  1. Create Namespace:

    kubectl create namespace ${namespace}

    A namespace is a virtual cluster backed by the same physical cluster. You can give it a name that is easy to memorize, such as the same name as cluster_name.

  2. Deploy the TiDB cluster:

    kubectl apply -f ${cluster_name} -n ${namespace}

    It is recommended to organize configurations for a TiDB cluster under a directory of cluster_name and save it as ${cluster_name}/tidb-cluster.yaml.

    If the server does not have an external network, you need to download the Docker image used by the TiDB cluster on a machine with Internet access and upload it to the server, and then use docker load to install the Docker image on the server.

    To deploy a TiDB cluster, you need the following Docker images (assuming the version of the TiDB cluster is v6.1.0):


    Next, download all these images with the following command:

    docker pull pingcap/pd:v6.1.0
    docker pull pingcap/tikv:v6.1.0
    docker pull pingcap/tidb:v6.1.0
    docker pull pingcap/tidb-binlog:v6.1.0
    docker pull pingcap/ticdc:v6.1.0
    docker pull pingcap/tiflash:v6.1.0
    docker pull pingcap/tidb-monitor-reloader:v1.0.1
    docker pull pingcap/tidb-monitor-initializer:v6.1.0
    docker pull grafana/grafana:6.0.1
    docker pull prom/prometheus:v2.18.1
    docker pull busybox:1.26.2
    docker save -o pd-v6.1.0.tar pingcap/pd:v6.1.0
    docker save -o tikv-v6.1.0.tar pingcap/tikv:v6.1.0
    docker save -o tidb-v6.1.0.tar pingcap/tidb:v6.1.0
    docker save -o tidb-binlog-v6.1.0.tar pingcap/tidb-binlog:v6.1.0
    docker save -o ticdc-v6.1.0.tar pingcap/ticdc:v6.1.0
    docker save -o tiflash-v6.1.0.tar pingcap/tiflash:v6.1.0
    docker save -o tidb-monitor-reloader-v1.0.1.tar pingcap/tidb-monitor-reloader:v1.0.1
    docker save -o tidb-monitor-initializer-v6.1.0.tar pingcap/tidb-monitor-initializer:v6.1.0
    docker save -o grafana-6.0.1.tar grafana/grafana:6.0.1
    docker save -o prometheus-v2.18.1.tar prom/prometheus:v2.18.1
    docker save -o busybox-1.26.2.tar busybox:1.26.2

    Next, upload these Docker images to the server, and execute docker load to install these Docker images on the server:

    docker load -i pd-v6.1.0.tar
    docker load -i tikv-v6.1.0.tar
    docker load -i tidb-v6.1.0.tar
    docker load -i tidb-binlog-v6.1.0.tar
    docker load -i ticdc-v6.1.0.tar
    docker load -i tiflash-v6.1.0.tar
    docker load -i tidb-monitor-reloader-v1.0.1.tar
    docker load -i tidb-monitor-initializer-v6.1.0.tar
    docker load -i grafana-6.0.1.tar
    docker load -i prometheus-v2.18.1.tar
    docker load -i busybox-1.26.2.tar
  3. View the Pod status:

    kubectl get po -n ${namespace} -l${cluster_name}

You can use TiDB Operator to deploy and manage multiple TiDB clusters in a single Kubernetes cluster by repeating the above procedure and replacing cluster_name with a different name.

Different clusters can be in the same or different namespace, which is based on your actual needs.


If you need to deploy a TiDB cluster on ARM64 machines, refer to Deploy a TiDB Cluster on ARM64 Machines.

Initialize the TiDB cluster

If you want to initialize your cluster after deployment, refer to Initialize a TiDB Cluster in Kubernetes.


By default, TiDB (starting from v4.0.2) periodically shares usage details with PingCAP to help understand how to improve the product. For details about what is shared and how to disable the sharing, see Telemetry.

Was this page helpful?