# tidb-docker-compose **Repository Path**: fellowfun110/tidb-docker-compose ## Basic Information - **Project Name**: tidb-docker-compose - **Description**: tidb-docker-compose - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 1 - **Created**: 2024-04-12 - **Last Updated**: 2024-04-12 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # TiDB docker-compose [![Build Status](https://travis-ci.org/pingcap/tidb-docker-compose.svg?branch=master)](https://travis-ci.org/pingcap/tidb-docker-compose) **WARNING: This is for testing only, DO NOT USE IN PRODUCTION!** ## Requirements * Docker >= 17.03 * Docker Compose >= 1.6.0 > **Note:** [Legacy Docker Toolbox](https://docs.docker.com/toolbox/toolbox_install_mac/) users must migrate to [Docker for Mac](https://store.docker.com/editions/community/docker-ce-desktop-mac), since it is tested that tidb-docker-compose cannot be started on Docker Toolbox and Docker Machine. > **Note:** It is recommended to disable SELinux. ## Quick start ```bash $ git clone https://github.com/pingcap/tidb-docker-compose.git $ cd tidb-docker-compose && docker-compose pull # Get the latest Docker images $ sudo setenforce 0 # Only on Linux $ docker-compose up -d $ mysql -h 127.0.0.1 -P 4000 -u root ``` * Access monitor at http://localhost:3000 (login with admin/admin if you want to modify grafana) * Access [tidb-vision](https://github.com/pingcap/tidb-vision) at http://localhost:8010 * Access Spark Web UI at http://localhost:8080 and access [TiSpark](https://github.com/pingcap/tispark) through spark://127.0.0.1:7077 ## Docker Swarm You can also use Docker Swarm to deploy a TiDB Platform cluster, and then you can scale the service using `docker stack` commands. ```bash $ docker swarm init # if your docker daemon is not already part of a swarm $ mkdir -p data logs $ docker stack deploy tidb -c docker-swarm.yml $ mysql -h 127.0.0.1 -P 4000 -u root ``` After deploying the stack, you can scale the number of TiDB Server instances in the cluster like this: ```bash $ docker service scale tidb_tidb=2 ``` Docker Swarm automatically load-balances across the containers that implement a scaled service, which you can see if you execute `select @@hostname` several times: ```bash $ mysql -h 127.0.0.1 -P 4000 -u root -te 'select @@hostname' +--------------+ | @@hostname | +--------------+ | 340092e0ec9e | +--------------+ $ mysql -h 127.0.0.1 -P 4000 -u root -te 'select @@hostname' +--------------+ | @@hostname | +--------------+ | e6f05ffe6274 | +--------------+ $ mysql -h 127.0.0.1 -P 4000 -u root -te 'select @@hostname' +--------------+ | @@hostname | +--------------+ | 340092e0ec9e | +--------------+ ``` If you want to connect to specific backend instances, for example to test concurrency by ensuring that you are connecting to distinct instances of tidb-server, you can use the `docker service ps` command to assemble a hostname for each container: ```bash $ docker service ps --no-trunc --format '{{.Name}}.{{.ID}}' tidb_tidb tidb_tidb.1.x3sc2sd66a88phsj103ohr6qq tidb_tidb.2.lk53apndq394cega46at853zw ``` To be able to resolve those hostnames, it's easiest to run the MySQL client in a container that has access to the swarm network: ```bash $ docker run --rm --network=tidb_default arey/mysql-client -h tidb_tidb.1.x3sc2sd66a88phsj103ohr6qq -P 4000 -u root -t -e 'select @@version' +-----------------------------------------+ | @@version | +-----------------------------------------+ | 5.7.25-TiDB-v3.0.0-beta.1-40-g873d9514b | +-----------------------------------------+ ``` To loop through all instances of TiDB Server, you can use a bash loop like this: ```bash for host in $(docker service ps --no-trunc --format '{{.Name}}.{{.ID}}' tidb_tidb) do docker run --rm --network tidb_default arey/mysql-client \ -h "$host" -P 4000 -u root -te "select @@hostname" done ``` To stop all services and remove all containers in the TiDB stack, execute `docker stack rm tidb`. ## Customize TiDB Cluster ### Configuration * config/pd.toml is copied from [PD repo](https://github.com/pingcap/pd/tree/master/conf) * config/tikv.toml is copied from [TiKV repo](https://github.com/pingcap/tikv/tree/master/etc) * config/tidb.toml is copied from [TiDB repo](https://github.com/pingcap/tidb/tree/master/config) * config/pump.toml is copied from [TiDB-Binlog repo](https://github.com/pingcap/tidb-binlog/tree/master/cmd/pump) * config/drainer.toml is copied from [TiDB-Binlog repo](https://github.com/pingcap/tidb-binlog/tree/master/cmd/drainer) If you find these configuration files outdated or mismatch with TiDB version, you can copy these files from their upstream repos and change their metrics addr with `pushgateway:9091`. Also `max-open-files` are configured to `1024` in tikv.toml to simplify quick start on Linux, because setting up ulimit on Linux with docker is quite tedious. And config/\*-dashboard.json are copied from [TiDB-Ansible repo](https://github.com/pingcap/tidb-ansible/tree/master/scripts) You can customize TiDB cluster configuration by editing docker-compose.yml and the above config files if you know what you're doing. But edit these files manually is tedious and error-prone, a template engine is strongly recommended. See the following steps ### Install Helm [Helm](https://helm.sh) is used as a template render engine ``` curl https://raw.githubusercontent.com/kubernetes/helm/master/scripts/get | bash ``` Or if you use Mac, you can use homebrew to install Helm by `brew install kubernetes-helm` ### Bring up TiDB cluster ```bash $ git clone https://github.com/pingcap/tidb-docker-compose.git $ cd tidb-docker-compose $ vi compose/values.yaml # custom cluster size, docker image, port mapping etc $ helm template compose > generated-docker-compose.yaml $ docker-compose -f generated-docker-compose.yaml pull # Get the latest Docker images $ docker-compose -f generated-docker-compose.yaml up -d # If you want to Bring up TiDB cluster with Binlog support $ vi compose/values.yaml # set tidb.enableBinlog to true $ helm template compose > generated-docker-compose-binlog.yaml $ docker-compose -f generated-docker-compose-binlog.yaml up -d # or you can use 'docker-compose-binlog.yml' file directly # Note: If the value of drainer.destDBType is "kafka" and # you want to consume the kafka messages outside the docker containers, # please update the kafka.advertisedHostName with your docker host IP in compose/values.yaml and # regenerate the 'generated-docker-compose-binlog.yaml' file ``` You can build docker image yourself for development test. * Build from binary For pd, tikv, tidb, pump and drainer comment their `image` and `buildPath` fields out. And then copy their binary files to pd/bin/pd-server, tikv/bin/tikv-server, tidb/bin/tidb-server, tidb-binlog/bin/pump and tidb-binlog/bin/drainer. These binary files can be built locally or downloaded from https://download.pingcap.org/tidb-latest-linux-amd64.tar.gz For tidbVision, comment its `image` and `buildPath` fields out. And then copy tidb-vision repo to tidb-vision/tidb-vision. * Build from source Leave pd, tikv, tidb and tidbVision `image` field empty and set their `buildPath` field to their source directory. For example, if your local tikv source directory is $GOPATH/src/github.com/pingcap/tikv, just set tikv `buildPath` to `$GOPATH/src/github.com/pingcap/tikv` *Note:* Compiling tikv from source consumes lots of memory, memory of Docker for Mac needs to be adjusted to greater than 6GB [tidb-vision](https://github.com/pingcap/tidb-vision) is a visiualization page of TiDB Cluster, it's WIP project and can be disabled by commenting `tidbVision` out. [TiSpark](https://github.com/pingcap/tispark) is a thin layer built for running Apache Spark on top of TiDB/TiKV to answer the complex OLAP queries. #### Host network mode (Linux) *Note:* Docker for Mac uses a Linux virtual machine, host network mode will not expose any services to host machine. So it's useless to use this mode. When using TiKV directly without TiDB, host network mode must be enabled. This way all services use host network without isolation. So you can access all services on the host machine. You can enable this mode by setting `networkMode: host` in compose/values.yaml and regenerate docker-compose.yml. When in this mode, prometheus address in configuration files should be changed from `prometheus:9090` to `127.0.0.1:9090`, and pushgateway address should be changed from `pushgateway:9091` to `127.0.0.1:9091`. These modification can be done by: ```bash # Note: this only needed when networkMode is `host` sed -i 's/pushgateway:9091/127.0.0.1:9091/g' config/* sed -i 's/prometheus:9090/127.0.0.1:9090/g' config/* ``` After all the above is done, you can start tidb-cluster as usual by `docker-compose -f generated-docker-compose.yml up -d` ### Debug TiDB/TiKV/PD instances Prerequisites: Pprof: This is a tool for visualization and analysis of profiling data. Follow [these instructions](https://github.com/google/pprof#building-pprof) to install pprof. Graphviz: [http://www.graphviz.org/](http://www.graphviz.org/), used to generate graphic visualizations of profiles. * debug TiDB or PD instances ```bash ### Use the following command to starts a web server for graphic visualizations of golang program profiles $ ./tool/container_debug -s pd0 -p /pd-server -w ``` The above command will produce graphic visualizations of profiles of `pd0` that can be accessed through the browser. * debug TiKV instances ```bash ### step 1: select a tikv instance(here is tikv0) and specify the binary path in container to enter debug container $ ./tool/container_debug -s tikv0 -p /tikv-server ### after step 1, we can generate flame graph for tikv0 in debug container $ ./run_flamegraph.sh 1 # 1 is the tikv0's process id ### also can fetch tikv0's stack informations with GDB in debug container $ gdb /tikv-server 1 -batch -ex "thread apply all bt" -ex "info threads" ``` ### Access TiDB cluster TiDB uses ports: 4000(mysql) and 10080(status) by default ```bash $ mysql -h 127.0.0.1 -P 4000 -u root --comments ``` And Grafana uses port 3000 by default, so open your browser at http://localhost:3000 to view monitor dashboard If you enabled tidb-vision, you can view it at http://localhost:8010 ### Access Spark shell and load TiSpark Insert some sample data to the TiDB cluster: ```bash $ docker-compose exec tispark-master bash $ cd /opt/spark/data/tispark-sample-data $ mysql --local-infile=1 -h tidb -P 4000 -u root --comments < dss.ddl ``` After the sample data is loaded into the TiDB cluster, you can access Spark Shell by `docker-compose exec tispark-master /opt/spark/bin/spark-shell`. ```bash $ docker-compose exec tispark-master /opt/spark/bin/spark-shell ... Spark context available as 'sc' (master = local[*], app id = local-1527045927617). Spark session available as 'spark'. Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 2.1.1 /_/ Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_172) Type in expressions to have them evaluated. Type :help for more information. scala> import org.apache.spark.sql.TiContext ... scala> val ti = new TiContext(spark) ... scala> ti.tidbMapDatabase("TPCH_001") ... scala> spark.sql("select count(*) from lineitem").show +--------+ |count(1)| +--------+ | 60175| +--------+ ``` You can also access Spark with Python or R using the following commands: ``` docker-compose exec tispark-master /opt/spark/bin/pyspark docker-compose exec tispark-master /opt/spark/bin/sparkR ``` More documents about TiSpark can be found [here](https://github.com/pingcap/tispark).