TiDB 6.1.0 Release Notes
Release date: June 13, 2022
TiDB version: 6.1.0
Quick access: Quick start | Production deployment
In 6.1.0, the key new features or improvements are as follows:
- List partitioning and list COLUMNS partitioning become GA, compatible with MySQL 5.7
- TiFlash partitioned table (dynamic pruning) becomes GA
- Support user-level lock management, compatible with MySQL
- Support non-transactional DML statements (only support
DELETE
) - TiFlash supports on-demand data compaction
- MPP introduces the window function framework
- TiCDC supports replicating changelogs to Kafka via Avro
- TiCDC supports splitting large transactions during replication, which significantly reduces replication latency caused by large transactions
- The optimistic mode for merging and migrating sharded tables becomes GA
New Features
SQL
List partitioning and list COLUMNS partitioning become GA. Both are compatible with MySQL 5.7.
User documents: List partitioning, List COLUMNS partitioning
TiFlash supports initiating a compact command. (experimental)
TiFlash v6.1.0 introduces the
ALTER TABLE ... COMPACT
statement, which provides a manual way to compact physical data based on the existing background compaction mechanism. With this statement, you can update data in earlier formats and improve read/write performance any time as appropriate. It is recommended that you execute this statement to compact data after upgrading your cluster to v6.1.0. This statement is an extension of the standard SQL syntax and therefore is compatible with MySQL clients. For scenarios other than TiFlash upgrade, usually there is no need to use this statement.TiFlash implements the window function framework and supports the following window functions:
RANK()
DENSE_RANK()
ROW_NUMBER()
Observability
Continuous Profiling supports the ARM architecture and TiFlash.
Grafana adds a Performance Overview dashboard to provide a system-level entry for overall performance diagnosis.
As a new dashboard in the TiDB visualized monitoring component Grafana, Performance Overview provides a system-level entry for overall performance diagnosis. According to the top-down performance analysis methodology, the Performance Overview dashboard reorganizes TiDB performance metrics based on database time breakdown and displays these metrics in different colors. By checking these colors, you can identify performance bottlenecks of the entire system at the first glance, which significantly reduces performance diagnosis time and simplifies performance analysis and diagnosis.
Performance
Support customized Region size
Starting from v6.1.0, you can configure
coprocessor.region-split-size
to set Regions to a larger size. This can effectively reduce the number of Regions, make Regions easier to manage, and improve the cluster performance and stability.Support using buckets to increase concurrency (experimental)
To help you further improve the query concurrency after setting Regions to a larger size, TiDB introduces the concept of bucket, which is a smaller range within a Region. Using buckets as the query unit can optimize concurrent query performance when Regions are set to a larger size. Using buckets as the query unit can also dynamically adjust the sizes of hotspot Regions to ensure the scheduling efficiency and load balance. This feature is currently experimental. It is not recommended to use it in production environments.
Use Raft Engine as the default log storage engine
Since v6.1.0, TiDB uses Raft Engine as the default storage engine for logs. Compared with RocksDB, Raft Engine can reduce TiKV I/O write traffic by up to 40% and CPU usage by 10%, while improving foreground throughput by about 5% and reducing tail latency by 20% under certain loads.
Support the join order hint syntax
The
LEADING
hint reminds the optimizer to use the specified order as the prefix of join operations. A good prefix of join can quickly reduce the amount of data at the early phase of join and improve the query performance.The
STRAIGHT_JOIN
hint reminds the optimizer to join tables in an order that is consistent with the order of tables in theFROM
clause.This provides a method for you to fix the order of table joins. A proper use of the hints can effectively enhance the SQL performance and cluster stability.
User document:
LEADING
,STRAIGHT_JOIN
, #29932
TiFlash supports four more functions:
FROM_DAYS
TO_DAYS
TO_SECONDS
WEEKOFYEAR
TiFlash supports partitioned tables in dynamic pruning mode.
To enhance performance in OLAP scenarios, dynamic pruning mode is supported for partitioned tables. If your TiDB is upgraded from versions earlier than v6.0.0, it is recommended that you manually update statistics of existing partitioned tables, so as to maximize the performance (not required for new installations or new partitions created after upgrade to v6.1.0).
User documents: Access partitioned tables in the MPP mode, Dynamic pruning mode, #3873
Stability
Automatic recovery from SST corruption
When RocksDB detects a damaged SST file in the background, TiKV will try to schedule the affected Peer and recover its data using other replicas. You can set the maximum allowable time for the recovery using the
background-error-recovery-window
parameter. If the recovery operation is not completed within the time window, TiKV will panic. This feature automatically detects and recovers recoverable damaged storage, thus improving the cluster stability.Support non-transactional DML statement
In the scenarios of large data processing, a single SQL statement with a large transaction might have a negative impact on the cluster stability and performance. Since v6.1.0, TiDB supports providing a syntax in which a
DELETE
statement is split into multiple statements for batch processing. The split statements compromise transactional atomicity and isolation but greatly improve the cluster stability. For detailed syntax, seeBATCH
.TiDB supports configuring the maximum GC wait time
The transaction of TiDB adopts the Multi-Version Concurrency Control (MVCC) mechanism. When the newly written data overwrites the old data, the old data is not replaced, and both versions of data are stored. The old data is cleaned up by the Garbage Collection (GC) task periodically, which helps reclaim storage space to improve the performance and stability of the cluster. GC is triggered every 10 minutes by default. To ensure that long-running transactions can access the corresponding historical data, when there are transactions in execution, the GC task is delayed. To ensure that the GC task is not delayed indefinitely, TiDB introduces the system variable
tidb_gc_max_wait_time
to control the maximum delay time of the GC task. If the maximum delay time is exceeded, the GC will be forcibly executed. The default value of the variable is 24 hours. This feature enables you to control the relationship between the GC waiting time and the long-running transaction, which improves the stability of the cluster.TiDB supports configuring the maximum execution time for automatic statistics collection tasks
Databases can effectively understand the distribution of data by collecting statistics, which helps generate reasonable execution plans and improve the efficiency of SQL execution. TiDB regularly collects statistics on frequently changed data objects in the background. However, collecting statistics takes up cluster resources and might affect the stable operation of the business during business peaks.
Starting from v6.1.0, TiDB introduces
tidb_max_auto_analyze_time
to control the maximum execution time for background statistics collection, which is 12 hours by default. When the application does not encounter a resource bottleneck, it is recommended not to modify this variable so that TiDB can timely collect statistics.
Ease of use
Support a one-stop online data recovery when multiple replicas are lost
Before TiDB v6.1.0, when multiple Region replicas are lost because of machine failure, users have to stop all TiKV servers and use TiKV Control to recover TiKV one by one. Since TiDB v6.1.0, the recovery process is fully automated, does not require to stop TiKV, and does not affect other applications online. The recovery process can be triggered using PD Control and provides a more user-friendly summary information.
Support viewing history statistics collection tasks
You can use the
SHOW ANALYZE STATUS
statement to show cluster-level statistics collection tasks. Before TiDB v6.1.0, theSHOW ANALYZE STATUS
statement shows instance-level tasks only, and history task records are cleared after a TiDB restart. Therefore, you cannot view history statistics collection time and details. Starting from TiDB v6.1.0, history records of statistics collection tasks are persisted and can be queried after a cluster restart, which provides a reference for troubleshooting query performance issues caused by statistics anomalies.Support modifying TiDB, TiKV, and TiFlash configurations dynamically
In earlier TiDB versions, after modifying a configuration item, you must restart the cluster to make the modification effective. This might interrupt online services. To address this issue, TiDB v6.1.0 introduces the dynamic configuration feature, which allows you to validate a parameter change without restarting the cluster. The specific optimizations are as follows:
- Transform some TiDB configuration items to system variables, so that they can be modified dynamically and persisted. Note that the original configuration items are deprecated after transformation. For a detailed list of the transformed configuration items, see Configuration file parameters.
- Support configuring some TiKV parameters online. For a detailed list of the parameters, see Others.
- Transform the TiFlash configuration item
max_threads
to a system variabletidb_max_tiflash_threads
, so that the configuration can be modified dynamically and persisted. Note that the original configuration item remains after transformation.
For v6.1.0 clusters upgraded (including online and offline upgrades) from earlier versions, note that:
- If the configuration items specified in the configuration file before the upgrade already exist, TiDB will automatically update the values of the configured items to those of the corresponding system variables during the upgrade process. In this way, after the upgrade, the system behavior is not affected by parameter optimization.
- The automatic update mentioned above occurs only once during the upgrade. After the upgrade, the deprecated configuration items are no longer effective.
This feature allows you to modify parameters dynamically, and validate and persist them, instead of restarting the system and interrupting services. This makes your daily maintenance easier.
Support killing queries or connections globally
You can control the Global Kill feature using the
enable-global-kill
configuration (enabled by default).Before TiDB v6.1.0, when an operation consumes a lot of resources and causes cluster stability issues, you have to connect to the target TiDB instance and then run the
KILL TIDB ${id};
command to terminate the target connection and operation. In the case of many TiDB instances, this method is not easy to use and prone to wrong operations. Starting from v6.1.0, theenable-global-kill
configuration is introduced and enabled by default. You can run the kill command in any TiDB instance to terminate a specified connection and operation, without worrying about incorrectly terminating other queries or sessions by mistake when there is a proxy between the client and TiDB. Currently, TiDB does not support using Ctrl+C to terminate queries or sessions.TiKV API V2 (experimental)
Before v6.1.0, when TiKV is used as Raw Key Value storage, TiKV only provides basic Key Value read and write capability because it only stores the raw data passed in by the client.
TiKV API V2 provides a new Raw Key Value storage format and access interface, including:
- The data is stored in MVCC and the change timestamp of the data is recorded. This feature will lay the foundation for implementing Change Data Capture and incremental backup and restore.
- Data is scoped according to different usage and supports co-existence of a single TiDB cluster, Transactional KV, RawKV applications.
MySQL compatibility
Support compatibility with user-level lock management with MySQL
User-level locks are a user-named lock management system provided by MySQL through built-in functions. The locking functions can provide lock blocking, waiting, and other lock management capabilities. User-level locks are also widely used in ORM frameworks, such as Rails, Elixir, and Ecto. Since v6.1.0, TiDB has supported MySQL-compatible user-level lock management, and supports
GET_LOCK
,RELEASE_LOCK
, andRELEASE_ALL_LOCKS
functions.
Data migration
The optimistic mode for merging and migrating sharded tables becomes GA
DM adds a large number of scenario tests for tasks that merge and migrate data from sharded tables in the optimistic mode, which covers 90% of the daily use scenarios. Compared with the pessimistic mode, the optimistic mode is simpler and more efficient to use. It is recommended to use the optimistic mode preferably after you are familiar with the usage notes.
DM WebUI supports starting a task according to the specified parameters
When starting a migration task, you can specify a start time and a safe mode duration. This is especially useful when you create an incremental migration task with lots of sources, eliminating the need to specify the binlog start position specifically for each source.
TiDB data share subscription
TiDB supports data sharing with various third-party data ecosystems
TiCDC supports sending TiDB incremental data to Kafka in the Avro format, allowing data sharing with third-parties, such as KSQL and Snowflake via Confluent.
TiCDC supports dispatching incremental data from TiDB to different Kafka topics by table, which, combined with the Canal-json format, allows sharing data directly with Flink.
TiCDC supports SASL GSSAPI authentication types and adds SASL authentication examples using Kafka.
TiCDC supports replicating
charset=GBK
tables.
Compatibility changes
System variables
Variable name | Change type | Description |
---|---|---|
tidb_enable_list_partition | Modified | The default value is changed from OFF to ON . |
tidb_mem_quota_query | Modified | This variable adds the GLOBAL scope, and the variable value persists to the cluster. |
tidb_query_log_max_len | Modified | The variable scope is changed from INSTANCE to GLOBAL. The variable value persists to the cluster, and the value range is changed to [0, 1073741824] . |
require_secure_transport | Newly added | This setting was previously a tidb.toml option (security.require-secure-transport ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_committer_concurrency | Newly added | This setting was previously a tidb.toml option (performance.committer-concurrency ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_enable_auto_analyze | Newly added | This setting was previously a tidb.toml option (run-auto-analyze ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_enable_new_only_full_group_by_check | Newly added | This variable controls the behavior when TiDB performs the ONLY_FULL_GROUP_BY check. |
tidb_enable_outer_join_reorder | Newly added | Since v6.1.0, the Join Reorder algorithm of TiDB supports Outer Join. This variable controls the support behavior, and the default value is ON . |
tidb_enable_prepared_plan_cache | Newly added | This setting was previously a tidb.toml option (prepared-plan-cache.enabled ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_gc_max_wait_time | Newly added | This variable is used to set the maximum time of GC safe point blocked by uncommitted transactions. |
tidb_max_auto_analyze_time | Newly added | This variable is used to specify the maximum execution time of auto analyze. |
tidb_max_tiflash_threads | Newly added | This variable is used to set the maximum concurrency for TiFlash to execute a request. |
tidb_mem_oom_action | Newly added | This setting was previously a tidb.toml option (oom-action ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_mem_quota_analyze | Newly added | This variable controls the maximum memory usage when TiDB updates statistics, including manually executed ANALYZE TABLE by users and automatic analyze tasks in the TiDB background. |
tidb_nontransactional_ignore_error | Newly added | This variable specifies whether to return error immediately when an error occurs in a non-transactional DML statement. |
tidb_prepared_plan_cache_memory_guard_ratio | Newly added | This setting was previously a tidb.toml option (prepared-plan-cache.memory-guard-ratio ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_prepared_plan_cache_size | Newly added | This setting was previously a tidb.toml option (prepared-plan-cache.capacity ), but changed to a system variable starting from TiDB v6.1.0. |
tidb_stats_cache_mem_quota | Newly added | This variable sets the memory quota for the TiDB statistics cache. |
Configuration file parameters
Configuration file | Configuration | Change type | Description |
---|---|---|---|
TiDB | committer-concurrency | Deleted | Replaced by the system variable tidb_committer_concurrency . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | lower-case-table-names | Deleted | Currently TiDB only supports lower_case_table_name=2 . If another value is set, after the cluster is upgraded to v6.1.0, the value is lost. |
TiDB | mem-quota-query | Deleted | Replaced by the system variable tidb_mem_quota_query . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | oom-action | Deleted | Replaced by the system variable tidb_mem_oom_action . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | prepared-plan-cache.capacity | Deleted | Replaced by the system variable tidb_prepared_plan_cache_size . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | prepared-plan-cache.enabled | Deleted | Replaced by the system variable tidb_enable_prepared_plan_cache . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | query-log-max-len | Deleted | Replaced by the system variable tidb_query_log_max_len . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | require-secure-transport | Deleted | Replaced by the system variable require_secure_transport . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | run-auto-analyze | Deleted | Replaced by the system variable tidb_enable_auto_analyze . This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable. |
TiDB | enable-global-kill | Newly added | Controls whether to enable the Global Kill (terminating queries or connections across instances) feature. When the value is true , both KILL and KILL TIDB statements can terminate queries or connections across instances so you do not need to worry about erroneously terminating queries or connections. |
TiDB | enable-stats-cache-mem-quota | Newly added | Controls whether to enable the memory quota for the statistics cache. |
TiKV | raft-engine.enable | Modified | The default value is changed from FALSE to TRUE . |
TiKV | region-max-keys | Modified | The default value is changed from 1440000 to region-split-keys / 2 * 3 . |
TiKV | region-max-size | Modified | The default value is changed from 144 MB to region-split-size / 2 * 3 . |
TiKV | coprocessor.enable-region-bucket | Newly added | Determines whether to divide a Region into smaller ranges called buckets. |
TiKV | coprocessor.region-bucket-size | Newly added | The size of a bucket when enable-region-bucket is true. |
TiKV | causal-ts.renew-batch-min-size | Newly added | The minimum number of locally cached timestamps. |
TiKV | causal-ts.renew-interval | Newly added | The interval at which the locally cached timestamps are refreshed. |
TiKV | max-snapshot-file-raw-size | Newly added | The snapshot file will split to multiple files when the snapshot file size exceeds this value. |
TiKV | raft-engine.memory-limit | Newly added | Specifies the limit on the memory usage of Raft Engine. |
TiKV | storage.background-error-recovery-window | Newly added | The maximum recovery time is allowed after RocksDB detects a recoverable background error. |
TiKV | storage.api-version | Newly added | The storage format and interface version used by TiKV when TiKV serves as the raw key-value store. |
PD | schedule.max-store-preparing-time | Newly added | Controls the maximum waiting time for the store to go online. |
TiCDC | enable-tls | Newly added | Whether to use TLS to connect to the downstream Kafka instance. |
TiCDC | sasl-gssapi-user sasl-gssapi-password sasl-gssapi-auth-type sasl-gssapi-service-name sasl-gssapi-realm sasl-gssapi-key-tab-path sasl-gssapi-kerberos-config-path | Newly added | Used to support SASL/GSSAPI authentication for Kafka. For details, see Configure sink URI with kafka . |
TiCDC | avro-decimal-handling-mode avro-bigint-unsigned-handling-mode | Newly added | Determines the output details of Avro format. |
TiCDC | dispatchers.topic | Newly added | Controls how TiCDC dispatches incremental data to different Kafka topics. |
TiCDC | dispatchers.partition | Newly added | dispatchers.partition is an alias for dispatchers.dispatcher . Controls how TiCDC dispatches incremental data to Kafka partitions. |
TiCDC | schema-registry | Newly added | Specifies the schema registry endpoint that stores Avro schema. |
DM | worker in the dmctl start-relay command | Deleted | This parameter is not recommended for use. Will provide a simpler implementation. |
DM | relay-dir in the source configuration file | Deleted | Replaced by the same configuration item in the worker configuration file. |
DM | is-sharding in the task configuration file | Deleted | Replaced by the shard-mode configuration item. |
DM | auto-fix-gtid in the task configuration file | Deleted | Deprecated in v5.x and officially deleted in v6.1.0. |
DM | meta-dir and charset in the source configuration file | Deleted | Deprecated in v5.x and officially deleted in v6.1.0. |
Others
Enable Prepared Plan Cache by default
Prepared Plan Cache is enabled by default in new clusters to cache the execution plans for
Prepare
/Execute
requests. In the subsequent execution, query plan optimization can be skipped and thus leads to a performance boost. Upgraded clusters inherit the configuration from the configuration file. New clusters use the new default values, which means Prepared Plan Cache is enabled by default and each session can cache 100 plans at most (capacity=100
). For the memory consumption of this feature, see memory management of Prepared Plan Cache.Prior to TiDB v6.1.0,
SHOW ANALYZE STATUS
shows instance-level tasks and the task records are cleared after TiDB restarts. Since TiDB v6.1.0,SHOW ANALYZE STATUS
shows cluster-level tasks, and the task records persist after the restart. Whentidb_analyze_version = 2
, theJob_info
column adds theanalyze option
information.Damaged SST files in TiKV might cause the TiKV process to panic. Before TiDB v6.1.0, damaged SST files caused TiKV to panic immediately. Since TiDB v6.1.0, the TiKV process will panic 1 hour after SST files are damaged.
The following TiKV configuration items support modifying values dynamically:
raftstore.raft-entry-max-size
quota.foreground-cpu-time
quota.foreground-write-bandwidth
quota.foreground-read-bandwidth
quota.max-delay-duration
server.grpc-memory-pool-quota
server.max-grpc-send-msg-len
server.raft-msg-max-batch-size
In v6.1.0, some configuration file parameters are converted to system variables. For v6.1.0 clusters upgraded (including online and offline upgrades ) from earlier versions, note that:
- If the configuration items specified in the configuration file before the upgrade already exist, TiDB will automatically update the values of the configured items to those of the corresponding system variables during the upgrade process. In this way, after the upgrade, the system behavior does not change thanks to parameter optimization.
- The automatic update mentioned above occurs only once during the upgrade. After the upgrade, the deprecated configuration items are no longer effective.
The Dashboard page is removed from DM WebUI.
When
dispatchers.topic
anddispatchers.partition
are enabled, TiCDC cannot be downgraded to versions earlier than v6.1.0.TiCDC Changefeed using the Avro protocol cannot be downgraded to versions earlier than v6.1.0.
Improvements
TiDB
- Improve the performance of the
UnionScanRead
operator #32433 - Improve the display of task types in the output of
EXPLAIN
(add the MPP task type) #33332 - Support using
rand()
as the default value of a column #10377 - Support using
uuid()
as the default value of a column #33870 - Support modifying the character set of columns from
latin1
toutf8
/utf8mb4
#34008
- Improve the performance of the
TiKV
- Improve the old value hit rate of CDC when using in-memory pessimistic lock #12279
- Improve the health check to detect unavailable Raftstore, so that the TiKV client can update Region Cache in time #12398
- Support setting memory limit on Raft Engine #12255
- TiKV automatically detects and deletes the damaged SST files to improve the product availability #10578
- CDC supports RawKV #11965
- Support splitting a large snapshot file into multiple files #11595
- Move the snapshot garbage collection from Raftstore to background thread to prevent snapshot GC from blocking Raftstore message loops #11966
- Support dynamic setting of the the maximum message length (
max-grpc-send-msg-len
) and the maximum batch size of gPRC messages (raft-msg-max-batch-size
) #12334 - Support executing online unsafe recovery plan through Raft #10483
PD
TiFlash
- Optimize memory calculation for an aggregate operator so that a more efficient algorithm is used in the merge phase #4451
Tools
Backup & Restore (BR)
- Support backing up and restoring empty databases #33866
TiDB Lightning
- Optimize Scatter Region to batch mode to improve the stability of the Scatter Region process #33618
TiCDC
- TiCDC supports splitting large transactions during replication, which significantly reduces replication latency caused by large transactions #5280
Bug fixes
TiDB
- Fix the issue of possible panic that might occur when the
in
function processes thebit
type data #33070 - Fix the issue of wrong query result because the
UnionScan
operator cannot maintain the order #33175 - Fix the issue that the Merge Join operator gets wrong results in certain cases #33042
- Fix the issue that the
index join
result might be wrong in the dynamic pruning mode #33231 - Fix the issue that data might not be garbage-collected when some partitions of a partitioned table is dropped #33620
- Fix the issue that some DDL statements might be stuck for a period after the PD node of a cluster is replaced #33908
- Fix the issue that the TiDB server might run out of memory when the
INFORMATION_SCHEMA.CLUSTER_SLOW_QUERY
table is queried. This issue can be triggered when you check slow queries on the Grafana dashboard #33893 - Fix the issue that the system variable
max_allowed_packet
does not take effect #31422 - Fix the issue of memory leak in the TopSQL module #34525 #34502
- Fix the issue that the Plan Cache might be wrong on the PointGet plan #32371
- Fix the issue that query result might be wrong when Plan Cache is started in the RC isolation level #34447
- Fix the issue of possible panic that might occur when the
TiKV
- Fix the issue that the Raft log lag is increasing when a TiKV instance is taken offline #12161
- Fix the issue that TiKV panics and destroys peers unexpectedly because the target Region to be merged is invalid #12232
- Fix the issue that TiKV reports the
failed to load_latest_options
error when upgrading from v5.3.1 or v5.4.0 to v6.0.0 or later versions #12269 - Fix the issue of OOM caused by appending Raft logs when the memory resource is insufficient #11379
- Fix the issue of TiKV panic caused by the race between destroying peers and batch splitting Regions #12368
- Fix the issue of TiKV memory usage spike in a short time after
stats_monitor
falls into a dead loop #12416 - Fix the issue that TiKV reports the
invalid store ID 0
error when using Follower Read #12478
PD
TiDB Dashboard
- Fix a bug that Top SQL cannot collect the CPU overhead of the SQL statements that were running before the Top SQL feature is enabled #33859
TiFlash
- Fix potential data inconsistency after a lot of INSERT and DELETE operations #4956
Tools
TiCDC
TiDB Data Migration (DM)
- Fix the
start-time
time zone issue and change DM behavior from using the downstream time zone to using the upstream time zone #5271 - Fix the issue that DM occupies more disk space after the task automatically resumes #3734 #5344
- Fix the problem that checkpoint flush may cause the data of failed rows to be skipped #5279
- Fix the issue that in some cases manually executing the filtered DDL in the downstream might cause task resumption failure #5272
- Fix an issue that the uppercase table cannot be replicated when
case-sensitive: true
is not set #5255 - Fix the DM worker panic issue that occurs when the primary key is not first in the index returned by the
SHOW CREATE TABLE
statement #5159 - Fix the issue that CPU usage may increase and a large amount of log is printed when GTID is enabled or when the task is automatically resumed #5063
- Fix the offline option and other usage issues in DM WebUI #4993
- Fix the issue that incremental tasks fail to start when GTID is empty in the upstream #3731
- Fix the issue that empty configurations may cause dm-master to panic #3732
- Fix the
TiDB Lightning
- Fix the issue that the precheck does not check local disk resources and cluster availability #34213
- Fix the issue of incorrect routing for schemas #33381
- Fix the issue that the PD configuration is not restored correctly when TiDB Lightning panics #31733
- Fix the issue of Local-backend import failure caused by out-of-bounds data in the
auto_increment
column #29737 - Fix the issue of local backend import failure when the
auto_random
orauto_increment
column is null #34208