Sign InTry Free

TiDB Docker Compose Deployment

This document describes how to quickly deploy a TiDB testing cluster with a single command using Docker Compose.

With Docker Compose, you can use a YAML file to configure application services in multiple containers. Then, with a single command, you can create and start all the services from your configuration.

Prerequisites

Make sure you have installed the following items on your machine:

Deploy TiDB using Docker Compose

  1. Download tidb-docker-compose.

    git clone https://github.com/pingcap/tidb-docker-compose.git
  2. Change the directory to tidb-docker-compose and get the latest TiDB Docker Images:

    cd tidb-docker-compose && docker-compose pull
  3. Start the TiDB cluster:

    docker-compose up -d
  4. Use the MySQL client to connect to TiDB to read and write data:

    mysql -h 127.0.0.1 -P 4000 -u root

Monitor the cluster

After you have successfully deployed a TiDB cluster, you can now monitor the TiDB cluster using one of the following methods:

  • Use Grafana to view the status of the cluster via http://localhost:3000 with the default account name and password: admin and admin.
  • Use TiDB-Vision, a cluster visualization tool, to see data transfer and load-balancing inside your cluster via http://localhost:8010.

Customize the cluster

After the deployment is completed, the following components are deployed by default:

  • 3 PD instances, 3 TiKV instances, 1 TiDB instance
  • Monitoring components: Prometheus, Pushgateway, Grafana
  • Data visualization component: tidb-vision

To customize the cluster, you can edit the docker-compose.yml file directly. It is recommended to generate docker-compose.yml using the Helm template engine, because manual editing is tedious and error-prone.

  1. Install Helm.

    Helm can be used as a template rendering engine. Make sure your Helm version >= 2.9.0 and < 3.0.0.

    To use Helm, you only need to download its binary file:

    curl https://raw.githubusercontent.com/kubernetes/helm/master/scripts/get | bash

    For macOS, you can also install Helm using the following command in Homebrew:

    brew install helm@2 brew link --force helm@2
  2. Download tidb-docker-compose.

    git clone https://github.com/pingcap/tidb-docker-compose.git
  3. Customize the cluster.

    cd tidb-docker-compose vim compose/values.yaml # custom the cluster size, docker image, port mapping and so on

    You can modify the configuration in values.yaml, such as the cluster size, TiDB image version, and so on.

    tidb-vision is the data visualization interface of the TiDB cluster, used to visually display the PD scheduling on TiKV data. If you do not need this component, comment out the tidbVision section.

    For PD, TiKV, TiDB and tidb-vision, you can build Docker images from GitHub source code or local files for development and testing.

    • To build PD, TiKV or TiDB images from the locally compiled binary file, you need to comment out the image field and copy the compiled binary file to the corresponding pd/bin/pd-server, tikv/bin/tikv-server, tidb/bin/tidb-server.
    • To build the tidb-vision image from local, you need to comment out the image field and copy the tidb-vision project to tidb-vision/tidb-vision.
  4. Generate the docker-compose.yml file.

    helm template compose > generated-docker-compose.yml
  5. Create and start the cluster using the generated docker-compose.yml file.

    docker-compose -f generated-docker-compose.yml pull # Get the latest Docker images docker-compose -f generated-docker-compose.yml up -d
  6. Access the cluster.

    mysql -h 127.0.0.1 -P 4000 -u root

    Access the Grafana monitoring interface:

    If tidb-vision is enabled, you can access the cluster data visualization interface: http://localhost:8010.

Access the Spark shell and load TiSpark

Insert some sample data to the TiDB cluster:

$ docker-compose exec tispark-master bash $ cd /opt/spark/data/tispark-sample-data $ mysql -h tidb -P 4000 -u root < dss.ddl

After the sample data is loaded into the TiDB cluster, you can access the Spark shell using docker-compose exec tispark-master /opt/spark/bin/spark-shell.

$ docker-compose exec tispark-master /opt/spark/bin/spark-shell ... Spark context available as 'sc' (master = local[*], app id = local-1527045927617). Spark session available as 'spark'. Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 2.1.1 /_/ Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_172) Type in expressions to have them evaluated. Type :help for more information. scala> import org.apache.spark.sql.TiContext ... scala> val ti = new TiContext(spark) ... scala> ti.tidbMapDatabase("TPCH_001") ... scala> spark.sql("select count(*) from lineitem").show +--------+ |count(1)| +--------+ | 60175| +--------+

You can also access Spark with Python or R using the following commands:

docker-compose exec tispark-master /opt/spark/bin/pyspark docker-compose exec tispark-master /opt/spark/bin/sparkR

For more details about TiSpark, see TiSpark Quick Start Guide.

Here is a 5-minute tutorial for macOS users that shows how to spin up a standard TiDB cluster using Docker Compose on your local computer.

Was this page helpful?

Download PDFRequest docs changesAsk questions on Discord
Playground
New
One-stop & interactive experience of TiDB's capabilities WITHOUT registration.
Products
TiDB
TiDB Dedicated
TiDB Serverless
Pricing
Get Demo
Get Started
© 2024 PingCAP. All Rights Reserved.
Privacy Policy.