Best Practices for SaaS Multi-Tenant Scenarios

h

This document introduces best practices for TiDB in SaaS (Software as a Service) multi-tenant environments, especially in scenarios where the number of tables in a single cluster exceeds one million. By making reasonable configurations and choices, you can enable TiDB to run efficiently and stably in SaaS scenarios while reducing resource consumption and costs.

TiDB hardware recommendations

It is recommended to use high-memory TiDB instances. For example:

  • For one million tables, use 32 GiB or more memory.
  • For three million tables, use 64 GiB or more memory.

High-memory TiDB instances allocate more cache space for Infoschema, Statistics, and execution plan caches, thereby improving cache hit rates and consequently enhancing business performance. Larger memory also mitigates performance fluctuations and stability issues caused by TiDB GC.

Recommended hardware configurations for TiKV and PD are as follows:

  • TiKV: 8 vCPUs and 32 GiB or more memory.
  • PD: 8 CPUs and 16 GiB or more memory.

Control the number of Regions

If you need to create a large number of tables (for example, more than 100,000), it is recommended to set the TiDB configuration item split-table to false to reduce the number of Regions, thus alleviating memory pressure on TiKV.

Configure caches

  • Starting from TiDB v8.4.0, TiDB loads table information involved in SQL statements into the Infoschema cache on demand during SQL execution.

    • You can monitor the size and hit rate of the Infoschema cache by observing the Infoschema v2 Cache Size and Infoschema v2 Cache Operation sub-panels under the Schema Load panel in TiDB Dashboard.
    • You can use the tidb_schema_cache_size system variable to adjust the memory limit of the Infoschema cache to meet business needs. The size of the Infoschema cache is linearly related to the number of different tables involved in SQL execution. In actual tests, fully caching metadata for one million tables (each with four columns, one primary key, and one index) requires about 2.4 GiB of memory.
  • TiDB loads table statistics involved in SQL statements into the Statistics cache on demand during SQL execution.

    • You can monitor the size and hit rate of the Statistics cache by observing the Stats Cache Cost and Stats Cache OPS sub-panels under the Statistics & Plan Management panel in TiDB Dashboard.
    • You can use the tidb_stats_cache_mem_quota system variable to adjust the memory limit of the Statistics cache to meet business needs. In actual tests, executing simple SQL (using the IndexRangeScan operator) on 100,000 tables consumes about 3.96 GiB of memory in the Statistics cache.

Collect statistics

  • Starting from TiDB v8.4.0, TiDB introduces the tidb_auto_analyze_concurrency system variable to control the number of concurrent auto-analyze operations that can run in a TiDB cluster. In multi-table scenarios, you can increase this concurrency as needed to improve the throughput of automatic analysis. As the concurrency value increases, the throughput and the CPU usage of the TiDB Owner node increase linearly. In actual tests, using a concurrency value of 16 allows automatic analysis of 320 tables (each with 10,000 rows, 4 columns, and 1 index) within one minute, consuming one CPU core of the TiDB Owner node.
  • The tidb_auto_build_stats_concurrency and tidb_build_sampling_stats_concurrency system variables control the concurrency of TiDB statistics construction. You can adjust them based on your scenario:
    • For scenarios with many partitioned tables, prioritize increasing the value of tidb_auto_build_stats_concurrency.
    • For scenarios with many columns, prioritize increasing the value of tidb_build_sampling_stats_concurrency.
  • To avoid excessive resource usage, ensure that the product of tidb_auto_analyze_concurrency, tidb_auto_build_stats_concurrency, and tidb_build_sampling_stats_concurrency does not exceed the number of TiDB CPU cores.

Query system tables efficiently

When querying system tables, it is recommended to add filters such as TABLE_SCHEMA, TABLE_NAME, or TIDB_TABLE_ID to avoid scanning a large amount of irrelevant data. This improves query speed and reduces resource consumption.

For example, in a scenario with three million tables:

  • Executing the following SQL statement consumes about 8 GiB of memory.

    SELECT COUNT(*) FROM information_schema.tables;
  • Executing the following SQL statement takes about 20 minutes.

    SELECT COUNT(*) FROM information_schema.views;

By adding appropriate filter conditions to the preceding SQL statements, memory consumption becomes negligible, and query time is reduced to milliseconds.

Handle connection-intensive scenarios

In SaaS multi-tenant scenarios, each user usually connects to TiDB to operate data in their own tenant (database). To support a high number of connections:

  • Increase the TiDB configuration item token-limit (1000 by default) to support more concurrent requests.
  • The memory usage of TiDB is roughly linear with the number of connections. In actual tests, 200,000 idle connections increase TiDB memory usage by about 30 GiB. It is recommended to increase TiDB memory specifications based on actual connection numbers.
  • If you use PREPARED statements, each connection maintains a session-level Prepared Plan Cache. If the DEALLOCATE statement is not executed for a long time, the cache might accumulate too many plans, increasing memory usage. In actual tests, 400,000 execution plans involving IndexRangeScan consume approximately 5 GiB of memory. It is recommended to increase memory specifications accordingly.

Use stale read carefully

When you use Stale Read, an outdated schema version might trigger a full load of historical schemas, which can significantly impact performance. To mitigate this issue, increase the value of tidb_schema_version_cache_limit (for example, to 255).

Optimize BR backup and restore

  • When restoring a full backup with millions of tables, it is recommended to use high-memory BR instances. For example:
    • For one million tables, use BR instances with 32 GiB or more memory.
    • For three million tables, use BR instances with 64 GiB or more memory.
  • BR log backup and snapshot restore consume additional TiKV memory. It is recommended to use TiKV instances with 32 GiB or more memory.
  • Adjust BR configurations pitr-batch-count and pitr-concurrency as needed to improve log restore speed.

Import data with TiDB Lightning

When importing millions of tables using TiDB Lightning, follow these recommendations:

Was this page helpful?