Sign InTry Free

Deploy TiDB Operator on Kubernetes

This document describes how to deploy TiDB Operator on Kubernetes.

Prerequisites

Before deploying TiDB Operator, make sure the following items are installed on your machine:

Deploy the Kubernetes cluster

TiDB Operator runs in the Kubernetes cluster. You can refer to the document of how to set up Kubernetes to set up a Kubernetes cluster. Make sure that the Kubernetes version is v1.12 or higher. If you want to deploy a very simple Kubernetes cluster for testing purposes, consult the Get Started document.

For some public cloud environments, refer to the following documents:

TiDB Operator uses Persistent Volumes to persist the data of TiDB cluster (including the database, monitoring data, and backup data), so the Kubernetes cluster must provide at least one kind of persistent volumes.

It is recommended to enable RBAC in the Kubernetes cluster.

Install Helm

Refer to Use Helm to install Helm and configure it with the official PingCAP chart repository.

Deploy TiDB Operator

Create CRD

TiDB Operator uses Custom Resource Definition (CRD) to extend Kubernetes. Therefore, to use TiDB Operator, you must first create the TidbCluster CRD, which is a one-time job in your Kubernetes cluster.

kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.2/manifests/crd.yaml

If the server cannot access the Internet, you need to download the crd.yaml file on a machine with Internet access before installing:

wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.2/manifests/crd.yaml kubectl create -f ./crd.yaml

If the following message is displayed, the CRD installation is successful:

kubectl get crd
NAME CREATED AT backups.pingcap.com 2020-06-11T07:59:40Z backupschedules.pingcap.com 2020-06-11T07:59:41Z restores.pingcap.com 2020-06-11T07:59:40Z tidbclusterautoscalers.pingcap.com 2020-06-11T07:59:42Z tidbclusters.pingcap.com 2020-06-11T07:59:38Z tidbinitializers.pingcap.com 2020-06-11T07:59:42Z tidbmonitors.pingcap.com 2020-06-11T07:59:41Z

Customize TiDB Operator deployment

To deploy TiDB Operator quickly, you can refer to Deploy TiDB Operator. This section describes how to customize the deployment of TiDB Operator.

After creating CRDs in the step above, there are two methods to deploy TiDB Operator on your Kubernetes cluster: online and offline.

When you use TiDB Operator, tidb-scheduler is not mandatory. Refer to tidb-scheduler and default-scheduler to confirm whether you need to deploy tidb-scheduler. If you do not need tidb-scheduler, you can configure scheduler.create: false in the values.yaml file, so tidb-scheduler is not deployed.

Online deployment

  1. Get the values.yaml file of the tidb-operator chart you want to deploy:

    mkdir -p ${HOME}/tidb-operator && \ helm inspect values pingcap/tidb-operator --version=${chart_version} > ${HOME}/tidb-operator/values-tidb-operator.yaml
  2. Configure TiDB Operator

    TiDB Operator manages all TiDB clusters in the Kubernetes cluster by default. If you only need it to manage clusters in a specific namespace, you can set clusterScoped: false in values.yaml.

    You can modify other items such as limits, requests, and replicas as needed.

  3. Deploy TiDB Operator

    helm install tidb-operator pingcap/tidb-operator --namespace=tidb-admin --version=${chart_version} -f ${HOME}/tidb-operator/values-tidb-operator.yaml && \ kubectl get po -n tidb-admin -l app.kubernetes.io/name=tidb-operator
  4. Upgrade TiDB Operator

    If you need to upgrade the TiDB Operator, modify the ${HOME}/tidb-operator/values-tidb-operator.yaml file, and then execute the following command to upgrade:

    helm upgrade tidb-operator pingcap/tidb-operator --namespace=tidb-admin -f ${HOME}/tidb-operator/values-tidb-operator.yaml

Offline installation

If your server cannot access the Internet, install TiDB Operator offline by the following steps:

  1. Download the tidb-operator chart

    If the server has no access to the Internet, you cannot configure the Helm repository to install the TiDB Operator component and other applications. At this time, you need to download the chart file needed for cluster installation on a machine with Internet access, and then copy it to the server.

    Use the following command to download the tidb-operator chart file:

    wget http://charts.pingcap.org/tidb-operator-v1.5.2.tgz

    Copy the tidb-operator-v1.5.2.tgz file to the target server and extract it to the current directory:

    tar zxvf tidb-operator.v1.5.2.tgz
  2. Download the Docker images used by TiDB Operator

    If the server has no access to the Internet, you need to download all Docker images used by TiDB Operator on a machine with Internet access and upload them to the server, and then use docker load to install the Docker image on the server.

    The Docker images used by TiDB Operator are:

    pingcap/tidb-operator:v1.5.2 pingcap/tidb-backup-manager:v1.5.2 bitnami/kubectl:latest pingcap/advanced-statefulset:v0.3.3 k8s.gcr.io/kube-scheduler:v1.16.9

    Among them, k8s.gcr.io/kube-scheduler:v1.16.9 should be consistent with the version of your Kubernetes cluster. You do not need to download it separately.

    Next, download all these images using the following command:

    docker pull pingcap/tidb-operator:v1.5.2 docker pull pingcap/tidb-backup-manager:v1.5.2 docker pull bitnami/kubectl:latest docker pull pingcap/advanced-statefulset:v0.3.3 docker save -o tidb-operator-v1.5.2.tar pingcap/tidb-operator:v1.5.2 docker save -o tidb-backup-manager-v1.5.2.tar pingcap/tidb-backup-manager:v1.5.2 docker save -o bitnami-kubectl.tar bitnami/kubectl:latest docker save -o advanced-statefulset-v0.3.3.tar pingcap/advanced-statefulset:v0.3.3

    Next, upload these Docker images to the server, and execute docker load to install these Docker images on the server:

    docker load -i tidb-operator-v1.5.2.tar docker load -i tidb-backup-manager-v1.5.2.tar docker load -i bitnami-kubectl.tar docker load -i advanced-statefulset-v0.3.3.tar
  3. Configure TiDB Operator

    TiDB Operator embeds a kube-scheduler to implement a custom scheduler. If you need to deploy tidb-scheduler, modify the ./tidb-operator/values.yaml file to configure the Docker image's name and version of this built-in kube-scheduler component. For example, if kube-scheduler in your Kubernetes cluster uses the image k8s.gcr.io/kube-scheduler:v1.16.9, set ./tidb-operator/values.yaml as follows:

    ... scheduler: serviceAccount: tidb-scheduler logLevel: 2 replicas: 1 schedulerName: tidb-scheduler resources: limits: cpu: 250m memory: 150Mi requests: cpu: 80m memory: 50Mi kubeSchedulerImageName: k8s.gcr.io/kube-scheduler kubeSchedulerImageTag: v1.16.9 ...

    You can modify other items such as limits, requests, and replicas as needed.

  4. Install TiDB Operator

    Install TiDB Operator using the following command:

    helm install tidb-operator ./tidb-operator --namespace=tidb-admin
  5. Upgrade TiDB Operator

    If you need to upgrade TiDB Operator, modify the ./tidb-operator/values.yaml file, and then execute the following command to upgrade:

    helm upgrade tidb-operator ./tidb-operator --namespace=tidb-admin

Customize TiDB Operator

To customize TiDB Operator, modify ${HOME}/tidb-operator/values-tidb-operator.yaml. The rest sections of the document use values.yaml to refer to ${HOME}/tidb-operator/values-tidb-operator.yaml

TiDB Operator contains two components:

  • tidb-controller-manager
  • tidb-scheduler

These two components are stateless and deployed via Deployment. You can customize resource limit, request, and replicas in the values.yaml file.

After modifying values.yaml, run the following command to apply this modification:

helm upgrade tidb-operator pingcap/tidb-operator --version=${chart_version} --namespace=tidb-admin -f ${HOME}/tidb-operator/values-tidb-operator.yaml

Was this page helpful?

Download PDFRequest docs changesAsk questions on Discord
Playground
New
One-stop & interactive experience of TiDB's capabilities WITHOUT registration.
Products
TiDB
TiDB Dedicated
TiDB Serverless
Pricing
Get Demo
Get Started
© 2024 PingCAP. All Rights Reserved.
Privacy Policy.