TiDB | |
Developer: | PingCAP Inc. |
Released: | [1] |
Programming Language: | Go (TiDB), Rust (TiKV) |
Language: | English, Chinese |
Genre: | NewSQL |
License: | Apache 2.0 |
TiDB (/’taɪdiːbi:/, "Ti" stands for Titanium) is an open-source NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads.[2] Designed to be MySQL compatible, it is developed and supported primarily by PingCAP and licensed under Apache 2.0. It is also available as a paid product. TiDB drew its initial design inspiration from Google's Spanner and F1 papers.[3] [4] [5]
See all TiDB release notes.
TiDB can expand both SQL processing and storage capacity by adding new nodes.
TiDB acts like it is a MySQL 8.0 server to applications. A user can continue to use all of the existing MySQL client libraries.[6] Because TiDB's SQL processing layer is built from scratch, not a MySQL fork.[7]
TiDB internally shards a table into small range-based chunks that are referred to as "Regions".[8] Each Region defaults to approximately 100 MB in size, and TiDB uses a two-phase commit internally to ensure that regions are maintained in a transactionally consistent way.
TiDB is designed to work in the cloud. The storage layer of TiDB, called TiKV, became a Cloud Native Computing Foundation (CNCF) member project in August 2018, as a Sandbox level project,[9] and became an incubation-level hosted project in May 2019.[10] TiKV graduated from CNCF in September 2020.[11]
TiDB can support both online transaction processing (OLTP) and online analytical processing (OLAP) workloads. TiDB has two storage engines: TiKV, a rowstore, and TiFlash, a columnstore.
TiDB uses the Raft consensus algorithm[12] to ensure that data is available and replicated throughout storage in Raft groups. In the event of failure, a Raft group will automatically elect a new leader for the failed member, and self-heal the TiDB cluster.
TiDB can be deployed in a Kubernetes-enabled cloud environment by using TiDB Operator.[13] An Operator is a method of packaging, deploying, and managing a Kubernetes application. It is designed for running stateful workloads and was first introduced by CoreOS in 2016.[14] TiDB Operator[15] was originally developed by PingCAP and open-sourced in August, 2018.[16] TiDB Operator can be used to deploy TiDB on a laptop,[17] Google Cloud Platform’s Google Kubernetes Engine,[18] and Amazon Web Services’ Elastic Container Service for Kubernetes.[19]
TiDB 4.0 introduces TiUP, a cluster operation and maintenance tool. It helps users quickly install and configure a TiDB cluster with a few commands.[20]
TiDB can be deployed using Ansible by using a TiDB Ansible playbook (not recommended).[21]
Docker can be used to deploy TiDB in a containerized environment on multiple nodes and multiple machines, and Docker Compose can be used to deploy TiDB with a single command for testing purposes.[22]
TiDB has a series of open-source tools built around it to help with data replication and migration for existing MySQL and MariaDB users.
TiDB Data Migration (DM) is suited for replicating data from already sharded MySQL or MariaDB tables to TiDB.[23] A common use case of DM is to connect MySQL or MariaDB tables to TiDB, treating TiDB almost as a slave, then directly run analytical workloads on this TiDB cluster in near real-time.
Backup & Restore (BR) is a distributed backup and restore tool for TiDB cluster data.[24]
Dumpling is a data export tool that exports data stored in TiDB or MySQL. It lets users make logical full backups or full dumps from TiDB or MySQL.[25]
TiDB Lightning is a tool that supports high speed full-import of a large MySQL dump into a new TiDB cluster. This tool is used to populate an initially empty TiDB cluster with much data, in order to speed up testing or production migration. The import speed improvement is achieved by parsing SQL statements into key-value pairs, then directly generate Sorted String Table (SST) files to RocksDB.[26] [27]
TiCDC is a Change data capture tool which streams data from TiDB to other systems like Apache Kafka.
TiDB Binlog is a tool used to collect the logical changes made to a TiDB cluster. It is used to provide incremental backup and replication, either between two TiDB clusters, or from a TiDB cluster to another downstream platform.[28] [29]