Skip to content

Latest commit

 

History

History
438 lines (285 loc) · 41.3 KB

release-6.1.0.md

File metadata and controls

438 lines (285 loc) · 41.3 KB
title summary
TiDB 6.1.0 Release Notes
Learn about the new features, compatibility changes, improvements, and bug fixes in TiDB 6.1.0.

TiDB 6.1.0 Release Notes

Release date: June 13, 2022

TiDB version: 6.1.0

Quick access: Quick start | Production deployment

In 6.1.0, the key new features or improvements are as follows:

  • List partitioning and list COLUMNS partitioning become GA, compatible with MySQL 5.7
  • TiFlash partitioned table (dynamic pruning) becomes GA
  • Support user-level lock management, compatible with MySQL
  • Support non-transactional DML statements (only support DELETE)
  • TiFlash supports on-demand data compaction
  • MPP introduces the window function framework
  • TiCDC supports replicating changelogs to Kafka via Avro
  • TiCDC supports splitting large transactions during replication, which significantly reduces replication latency caused by large transactions
  • The optimistic mode for merging and migrating sharded tables becomes GA

New Features

SQL

  • List partitioning and list COLUMNS partitioning become GA. Both are compatible with MySQL 5.7.

    User documents: List partitioning, List COLUMNS partitioning

  • TiFlash supports initiating a compact command. (experimental)

    TiFlash v6.1.0 introduces the ALTER TABLE ... COMPACT statement, which provides a manual way to compact physical data based on the existing background compaction mechanism. With this statement, you can update data in earlier formats and improve read/write performance any time as appropriate. It is recommended that you execute this statement to compact data after upgrading your cluster to v6.1.0. This statement is an extension of the standard SQL syntax and therefore is compatible with MySQL clients. For scenarios other than TiFlash upgrade, usually there is no need to use this statement.

    User document, #4145

  • TiFlash implements the window function framework and supports the following window functions:

    • RANK()
    • DENSE_RANK()
    • ROW_NUMBER()

    User document, #33072

Observability

  • Continuous Profiling supports the ARM architecture and TiFlash.

    User document

  • Grafana adds a Performance Overview dashboard to provide a system-level entry for overall performance diagnosis.

    As a new dashboard in the TiDB visualized monitoring component Grafana, Performance Overview provides a system-level entry for overall performance diagnosis. According to the top-down performance analysis methodology, the Performance Overview dashboard reorganizes TiDB performance metrics based on database time breakdown and displays these metrics in different colors. By checking these colors, you can identify performance bottlenecks of the entire system at the first glance, which significantly reduces performance diagnosis time and simplifies performance analysis and diagnosis.

    User document

Performance

  • Support customized Region size

    Starting from v6.1.0, you can configure coprocessor.region-split-size to set Regions to a larger size. This can effectively reduce the number of Regions, make Regions easier to manage, and improve the cluster performance and stability.

    User document, #11515

  • Support using buckets to increase concurrency (experimental)

    To help you further improve the query concurrency after setting Regions to a larger size, TiDB introduces the concept of bucket, which is a smaller range within a Region. Using buckets as the query unit can optimize concurrent query performance when Regions are set to a larger size. Using buckets as the query unit can also dynamically adjust the sizes of hotspot Regions to ensure the scheduling efficiency and load balance. This feature is currently experimental. It is not recommended to use it in production environments.

    User document, #11515

  • Use Raft Engine as the default log storage engine

    Since v6.1.0, TiDB uses Raft Engine as the default storage engine for logs. Compared with RocksDB, Raft Engine can reduce TiKV I/O write traffic by up to 40% and CPU usage by 10%, while improving foreground throughput by about 5% and reducing tail latency by 20% under certain loads.

    User document, #95

  • Support the join order hint syntax

    • The LEADING hint reminds the optimizer to use the specified order as the prefix of join operations. A good prefix of join can quickly reduce the amount of data at the early phase of join and improve the query performance.
    • The STRAIGHT_JOIN hint reminds the optimizer to join tables in an order that is consistent with the order of tables in the FROM clause.

    This provides a method for you to fix the order of table joins. A proper use of the hints can effectively enhance the SQL performance and cluster stability.

    User document: LEADING, STRAIGHT_JOIN, #29932

  • TiFlash supports four more functions:

    • FROM_DAYS
    • TO_DAYS
    • TO_SECONDS
    • WEEKOFYEAR

    User document, #4679, #4678, #4677

  • TiFlash supports partitioned tables in dynamic pruning mode.

    To enhance performance in OLAP scenarios, dynamic pruning mode is supported for partitioned tables. If your TiDB is upgraded from versions earlier than v6.0.0, it is recommended that you manually update statistics of existing partitioned tables, so as to maximize the performance (not required for new installations or new partitions created after upgrade to v6.1.0).

    User documents: Access partitioned tables in the MPP mode, Dynamic pruning mode, #3873

Stability

  • Automatic recovery from SST corruption

    When RocksDB detects a damaged SST file in the background, TiKV will try to schedule the affected Peer and recover its data using other replicas. You can set the maximum allowable time for the recovery using the background-error-recovery-window parameter. If the recovery operation is not completed within the time window, TiKV will panic. This feature automatically detects and recovers recoverable damaged storage, thus improving the cluster stability.

    User document, #10578

  • Support non-transactional DML statement

    In the scenarios of large data processing, a single SQL statement with a large transaction might have a negative impact on the cluster stability and performance. Since v6.1.0, TiDB supports providing a syntax in which a DELETE statement is split into multiple statements for batch processing. The split statements compromise transactional atomicity and isolation but greatly improve the cluster stability. For detailed syntax, see BATCH.

    User document

  • TiDB supports configuring the maximum GC wait time

    The transaction of TiDB adopts the Multi-Version Concurrency Control (MVCC) mechanism. When the newly written data overwrites the old data, the old data is not replaced, and both versions of data are stored. The old data is cleaned up by the Garbage Collection (GC) task periodically, which helps reclaim storage space to improve the performance and stability of the cluster. GC is triggered every 10 minutes by default. To ensure that long-running transactions can access the corresponding historical data, when there are transactions in execution, the GC task is delayed. To ensure that the GC task is not delayed indefinitely, TiDB introduces the system variable tidb_gc_max_wait_time to control the maximum delay time of the GC task. If the maximum delay time is exceeded, the GC will be forcibly executed. The default value of the variable is 24 hours. This feature enables you to control the relationship between the GC waiting time and the long-running transaction, which improves the stability of the cluster.

    User document

  • TiDB supports configuring the maximum execution time for automatic statistics collection tasks

    Databases can effectively understand the distribution of data by collecting statistics, which helps generate reasonable execution plans and improve the efficiency of SQL execution. TiDB regularly collects statistics on frequently changed data objects in the background. However, collecting statistics takes up cluster resources and might affect the stable operation of the business during business peaks.

    Starting from v6.1.0, TiDB introduces tidb_max_auto_analyze_time to control the maximum execution time for background statistics collection, which is 12 hours by default. When the application does not encounter a resource bottleneck, it is recommended not to modify this variable so that TiDB can timely collect statistics.

    User document

Ease of use

  • Support a one-stop online data recovery when multiple replicas are lost

    Before TiDB v6.1.0, when multiple Region replicas are lost because of machine failure, users have to stop all TiKV servers and use TiKV Control to recover TiKV one by one. Since TiDB v6.1.0, the recovery process is fully automated, does not require to stop TiKV, and does not affect other applications online. The recovery process can be triggered using PD Control and provides a more user-friendly summary information.

    User document, #10483

  • Support viewing history statistics collection tasks

    You can use the SHOW ANALYZE STATUS statement to show cluster-level statistics collection tasks. Before TiDB v6.1.0, the SHOW ANALYZE STATUS statement shows instance-level tasks only, and history task records are cleared after a TiDB restart. Therefore, you cannot view history statistics collection time and details. Starting from TiDB v6.1.0, history records of statistics collection tasks are persisted and can be queried after a cluster restart, which provides a reference for troubleshooting query performance issues caused by statistics anomalies.

    User document

  • Support modifying TiDB, TiKV, and TiFlash configurations dynamically

    In earlier TiDB versions, after modifying a configuration item, you must restart the cluster to make the modification effective. This might interrupt online services. To address this issue, TiDB v6.1.0 introduces the dynamic configuration feature, which allows you to validate a parameter change without restarting the cluster. The specific optimizations are as follows:

    • Transform some TiDB configuration items to system variables, so that they can be modified dynamically and persisted. Note that the original configuration items are deprecated after transformation. For a detailed list of the transformed configuration items, see Configuration file parameters.
    • Support configuring some TiKV parameters online. For a detailed list of the parameters, see Others.
    • Transform the TiFlash configuration item max_threads to a system variable tidb_max_tiflash_threads, so that the configuration can be modified dynamically and persisted. Note that the original configuration item remains after transformation.

    For v6.1.0 clusters upgraded (including online and offline upgrades) from earlier versions, note that:

    • If the configuration items specified in the configuration file before the upgrade already exist, TiDB will automatically update the values of the configured items to those of the corresponding system variables during the upgrade process. In this way, after the upgrade, the system behavior is not affected by parameter optimization.
    • The automatic update mentioned above occurs only once during the upgrade. After the upgrade, the deprecated configuration items are no longer effective.

    This feature allows you to modify parameters dynamically, and validate and persist them, instead of restarting the system and interrupting services. This makes your daily maintenance easier.

    User document

  • Support killing queries or connections globally

    You can control the Global Kill feature using the enable-global-kill configuration (enabled by default).

    Before TiDB v6.1.0, when an operation consumes a lot of resources and causes cluster stability issues, you have to connect to the target TiDB instance and then run the KILL TIDB ${id}; command to terminate the target connection and operation. In the case of many TiDB instances, this method is not easy to use and prone to wrong operations. Starting from v6.1.0, the enable-global-kill configuration is introduced and enabled by default. You can run the kill command in any TiDB instance to terminate a specified connection and operation, without worrying about incorrectly terminating other queries or sessions by mistake when there is a proxy between the client and TiDB. Currently, TiDB does not support using Ctrl+C to terminate queries or sessions.

    User document, #8854

  • TiKV API V2 (experimental)

    Before v6.1.0, when TiKV is used as Raw Key Value storage, TiKV only provides basic Key Value read and write capability because it only stores the raw data passed in by the client.

    TiKV API V2 provides a new Raw Key Value storage format and access interface, including:

    • The data is stored in MVCC and the change timestamp of the data is recorded. This feature will lay the foundation for implementing Change Data Capture and incremental backup and restore.
    • Data is scoped according to different usage and supports co-existence of a single TiDB cluster, Transactional KV, RawKV applications.

    Due to significant changes in the underlying storage format, after enabling API V2, you cannot roll back a TiKV cluster to a version earlier than v6.1.0. Downgrading TiKV might result in data corruption.

    User document, #11745

MySQL compatibility

  • Support compatibility with user-level lock management with MySQL

    User-level locks are a user-named lock management system provided by MySQL through built-in functions. The locking functions can provide lock blocking, waiting, and other lock management capabilities. User-level locks are also widely used in ORM frameworks, such as Rails, Elixir, and Ecto. Since v6.1.0, TiDB has supported MySQL-compatible user-level lock management, and supports GET_LOCK, RELEASE_LOCK, and RELEASE_ALL_LOCKS functions.

    User document, #14994

Data migration

  • The optimistic mode for merging and migrating sharded tables becomes GA

    DM adds a large number of scenario tests for tasks that merge and migrate data from sharded tables in the optimistic mode, which covers 90% of the daily use scenarios. Compared with the pessimistic mode, the optimistic mode is simpler and more efficient to use. It is recommended to use the optimistic mode preferably after you are familiar with the usage notes.

    User document

  • DM WebUI supports starting a task according to the specified parameters

    When starting a migration task, you can specify a start time and a safe mode duration. This is especially useful when you create an incremental migration task with lots of sources, eliminating the need to specify the binlog start position specifically for each source.

    User document, #5442

TiDB data share subscription

  • TiDB supports data sharing with various third-party data ecosystems

    • TiCDC supports sending TiDB incremental data to Kafka in the Avro format, allowing data sharing with third-parties, such as KSQL and Snowflake via Confluent.

      User document, #5338

    • TiCDC supports dispatching incremental data from TiDB to different Kafka topics by table, which, combined with the Canal-json format, allows sharing data directly with Flink.

      User document, #4423

    • TiCDC supports SASL GSSAPI authentication types and adds SASL authentication examples using Kafka.

      User document, #4423

  • TiCDC supports replicating charset=GBK tables.

    User document, #4806

Compatibility changes

System variables

Variable name Change type Description
tidb_enable_list_partition Modified The default value is changed from OFF to ON.
tidb_mem_quota_query Modified This variable adds the GLOBAL scope, and the variable value persists to the cluster.
tidb_query_log_max_len Modified The variable scope is changed from INSTANCE to GLOBAL. The variable value persists to the cluster, and the value range is changed to [0, 1073741824].
require_secure_transport Newly added This setting was previously a tidb.toml option (security.require-secure-transport), but changed to a system variable starting from TiDB v6.1.0.
tidb_committer_concurrency Newly added This setting was previously a tidb.toml option (performance.committer-concurrency), but changed to a system variable starting from TiDB v6.1.0.
tidb_enable_auto_analyze Newly added This setting was previously a tidb.toml option (run-auto-analyze), but changed to a system variable starting from TiDB v6.1.0.
tidb_enable_new_only_full_group_by_check Newly added This variable controls the behavior when TiDB performs the ONLY_FULL_GROUP_BY check.
tidb_enable_outer_join_reorder Newly added Since v6.1.0, the Join Reorder algorithm of TiDB supports Outer Join. This variable controls the support behavior, and the default value is ON.
tidb_enable_prepared_plan_cache Newly added This setting was previously a tidb.toml option (prepared-plan-cache.enabled), but changed to a system variable starting from TiDB v6.1.0.
tidb_gc_max_wait_time Newly added This variable is used to set the maximum time of GC safe point blocked by uncommitted transactions.
tidb_max_auto_analyze_time Newly added This variable is used to specify the maximum execution time of auto analyze.
tidb_max_tiflash_threads Newly added This variable is used to set the maximum concurrency for TiFlash to execute a request.
tidb_mem_oom_action Newly added This setting was previously a tidb.toml option (oom-action), but changed to a system variable starting from TiDB v6.1.0.
tidb_mem_quota_analyze Newly added This variable controls the maximum memory usage when TiDB updates statistics, including manually executed ANALYZE TABLE by users and automatic analyze tasks in the TiDB background.
tidb_nontransactional_ignore_error Newly added This variable specifies whether to return error immediately when an error occurs in a non-transactional DML statement.
tidb_prepared_plan_cache_memory_guard_ratio Newly added This setting was previously a tidb.toml option (prepared-plan-cache.memory-guard-ratio), but changed to a system variable starting from TiDB v6.1.0.
tidb_prepared_plan_cache_size Newly added This setting was previously a tidb.toml option (prepared-plan-cache.capacity), but changed to a system variable starting from TiDB v6.1.0.
tidb_stats_cache_mem_quota Newly added This variable sets the memory quota for the TiDB statistics cache.

Configuration file parameters

Configuration file Configuration Change type Description
TiDB committer-concurrency Deleted Replaced by the system variable tidb_committer_concurrency. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB lower-case-table-names Deleted Currently TiDB only supports lower_case_table_name=2. If another value is set, after the cluster is upgraded to v6.1.0, the value is lost.
TiDB mem-quota-query Deleted Replaced by the system variable tidb_mem_quota_query. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB oom-action Deleted Replaced by the system variable tidb_mem_oom_action. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB prepared-plan-cache.capacity Deleted Replaced by the system variable tidb_prepared_plan_cache_size. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB prepared-plan-cache.enabled Deleted Replaced by the system variable tidb_enable_prepared_plan_cache. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB query-log-max-len Deleted Replaced by the system variable tidb_query_log_max_len. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB require-secure-transport Deleted Replaced by the system variable require_secure_transport. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB run-auto-analyze Deleted Replaced by the system variable tidb_enable_auto_analyze. This configuration item is no longer valid, if you want to modify the value, you need to modify the corresponding system variable.
TiDB enable-global-kill Newly added Controls whether to enable the Global Kill (terminating queries or connections across instances) feature. When the value is true, both KILL and KILL TIDB statements can terminate queries or connections across instances so you do not need to worry about erroneously terminating queries or connections.
TiDB enable-stats-cache-mem-quota Newly added Controls whether to enable the memory quota for the statistics cache.
TiKV raft-engine.enable Modified The default value is changed from FALSE to TRUE.
TiKV region-max-keys Modified The default value is changed from 1440000 to region-split-keys / 2 * 3.
TiKV region-max-size Modified The default value is changed from 144 MB to region-split-size / 2 * 3.
TiKV coprocessor.enable-region-bucket Newly added Determines whether to divide a Region into smaller ranges called buckets.
TiKV coprocessor.region-bucket-size Newly added The size of a bucket when enable-region-bucket is true.
TiKV causal-ts.renew-batch-min-size Newly added The minimum number of locally cached timestamps.
TiKV causal-ts.renew-interval Newly added The interval at which the locally cached timestamps are refreshed.
TiKV max-snapshot-file-raw-size Newly added The snapshot file will split to multiple files when the snapshot file size exceeds this value.
TiKV raft-engine.memory-limit Newly added Specifies the limit on the memory usage of Raft Engine.
TiKV storage.background-error-recovery-window Newly added The maximum recovery time is allowed after RocksDB detects a recoverable background error.
TiKV storage.api-version Newly added The storage format and interface version used by TiKV when TiKV serves as the raw key-value store.
PD schedule.max-store-preparing-time Newly added Controls the maximum waiting time for the store to go online.
TiCDC enable-tls Newly added Whether to use TLS to connect to the downstream Kafka instance.
TiCDC sasl-gssapi-user
sasl-gssapi-password
sasl-gssapi-auth-type
sasl-gssapi-service-name
sasl-gssapi-realm
sasl-gssapi-key-tab-path
sasl-gssapi-kerberos-config-path
Newly added Used to support SASL/GSSAPI authentication for Kafka. For details, see Configure sink URI with kafka.
TiCDC avro-decimal-handling-mode
avro-bigint-unsigned-handling-mode
Newly added Determines the output details of Avro format.
TiCDC dispatchers.topic Newly added Controls how TiCDC dispatches incremental data to different Kafka topics.
TiCDC dispatchers.partition Newly added dispatchers.partition is an alias for dispatchers.dispatcher. Controls how TiCDC dispatches incremental data to Kafka partitions.
TiCDC schema-registry Newly added Specifies the schema registry endpoint that stores Avro schema.
DM worker in the dmctl start-relay command Deleted This parameter is not recommended for use. Will provide a simpler implementation.
DM relay-dir in the source configuration file Deleted Replaced by the same configuration item in the worker configuration file.
DM is-sharding in the task configuration file Deleted Replaced by the shard-mode configuration item.
DM auto-fix-gtid in the task configuration file Deleted Deprecated in v5.x and officially deleted in v6.1.0.
DM meta-dir and charset in the source configuration file Deleted Deprecated in v5.x and officially deleted in v6.1.0.

Others

  • Enable Prepared Plan Cache by default

    Prepared Plan Cache is enabled by default in new clusters to cache the execution plans for Prepare / Execute requests. In the subsequent execution, query plan optimization can be skipped and thus leads to a performance boost. Upgraded clusters inherit the configuration from the configuration file. New clusters use the new default values, which means Prepared Plan Cache is enabled by default and each session can cache 100 plans at most (capacity=100). For the memory consumption of this feature, see memory management of Prepared Plan Cache.

  • Prior to TiDB v6.1.0, SHOW ANALYZE STATUS shows instance-level tasks and the task records are cleared after TiDB restarts. Since TiDB v6.1.0, SHOW ANALYZE STATUS shows cluster-level tasks, and the task records persist after the restart. When tidb_analyze_version = 2, the Job_info column adds the analyze option information.

  • Damaged SST files in TiKV might cause the TiKV process to panic. Before TiDB v6.1.0, damaged SST files caused TiKV to panic immediately. Since TiDB v6.1.0, the TiKV process will panic 1 hour after SST files are damaged.

  • The following TiKV configuration items support modifying values dynamically:

    • raftstore.raft-entry-max-size
    • quota.foreground-cpu-time
    • quota.foreground-write-bandwidth
    • quota.foreground-read-bandwidth
    • quota.max-delay-duration
    • server.grpc-memory-pool-quota
    • server.max-grpc-send-msg-len
    • server.raft-msg-max-batch-size
  • In v6.1.0, some configuration file parameters are converted to system variables. For v6.1.0 clusters upgraded (including online and offline upgrades ) from earlier versions, note that:

    • If the configuration items specified in the configuration file before the upgrade already exist, TiDB will automatically update the values of the configured items to those of the corresponding system variables during the upgrade process. In this way, after the upgrade, the system behavior does not change thanks to parameter optimization.
    • The automatic update mentioned above occurs only once during the upgrade. After the upgrade, the deprecated configuration items are no longer effective.
  • The Dashboard page is removed from DM WebUI.

  • When dispatchers.topic and dispatchers.partition are enabled, TiCDC cannot be downgraded to versions earlier than v6.1.0.

  • TiCDC Changefeed using the Avro protocol cannot be downgraded to versions earlier than v6.1.0.

Improvements

  • TiDB

    • Improve the performance of the UnionScanRead operator #32433
    • Improve the display of task types in the output of EXPLAIN (add the MPP task type) #33332
    • Support using rand() as the default value of a column #10377
    • Support using uuid() as the default value of a column #33870
    • Support modifying the character set of columns from latin1 to utf8/utf8mb4 #34008
  • TiKV

    • Improve the old value hit rate of CDC when using in-memory pessimistic lock #12279
    • Improve the health check to detect unavailable Raftstore, so that the TiKV client can update Region Cache in time #12398
    • Support setting memory limit on Raft Engine #12255
    • TiKV automatically detects and deletes the damaged SST files to improve the product availability #10578
    • CDC supports RawKV #11965
    • Support splitting a large snapshot file into multiple files #11595
    • Move the snapshot garbage collection from Raftstore to background thread to prevent snapshot GC from blocking Raftstore message loops #11966
    • Support dynamic setting of the maximum message length (max-grpc-send-msg-len) and the maximum batch size of gPRC messages (raft-msg-max-batch-size) #12334
    • Support executing online unsafe recovery plan through Raft #10483
  • PD

    • Support time-to-live (TTL) for region labels #4694
    • Support Region Buckets #4668
    • Disable compiling swagger server by default #4932
  • TiFlash

    • Optimize memory calculation for an aggregate operator so that a more efficient algorithm is used in the merge phase #4451
  • Tools

    • Backup & Restore (BR)

      • Support backing up and restoring empty databases #33866
    • TiDB Lightning

      • Optimize Scatter Region to batch mode to improve the stability of the Scatter Region process #33618
    • TiCDC

      • TiCDC supports splitting large transactions during replication, which significantly reduces replication latency caused by large transactions #5280

Bug fixes

  • TiDB

    • Fix the issue of possible panic that might occur when the in function processes the bit type data #33070
    • Fix the issue of wrong query result because the UnionScan operator cannot maintain the order #33175
    • Fix the issue that the Merge Join operator gets wrong results in certain cases #33042
    • Fix the issue that the index join result might be wrong in the dynamic pruning mode #33231
    • Fix the issue that data might not be garbage-collected when some partitions of a partitioned table is dropped #33620
    • Fix the issue that some DDL statements might be stuck for a period after the PD node of a cluster is replaced #33908
    • Fix the issue that the TiDB server might run out of memory when the INFORMATION_SCHEMA.CLUSTER_SLOW_QUERY table is queried. This issue can be triggered when you check slow queries on the Grafana dashboard #33893
    • Fix the issue that the system variable max_allowed_packet does not take effect #31422
    • Fix the issue of memory leak in the TopSQL module #34525 #34502
    • Fix the issue that the Plan Cache might be wrong on the PointGet plan #32371
    • Fix the issue that query result might be wrong when Plan Cache is started in the RC isolation level #34447
  • TiKV

    • Fix the issue that the Raft log lag is increasing when a TiKV instance is taken offline #12161
    • Fix the issue that TiKV panics and destroys peers unexpectedly because the target Region to be merged is invalid #12232
    • Fix the issue that TiKV reports the failed to load_latest_options error when upgrading from v5.3.1 or v5.4.0 to v6.0.0 or later versions #12269
    • Fix the issue of OOM caused by appending Raft logs when the memory resource is insufficient #11379
    • Fix the issue of TiKV panic caused by the race between destroying peers and batch splitting Regions #12368
    • Fix the issue of TiKV memory usage spike in a short time after stats_monitor falls into a dead loop #12416
    • Fix the issue that TiKV reports the invalid store ID 0 error when using Follower Read #12478
  • PD

    • Fix the wrong status code of not leader #4797
    • Fix a bug of TSO fallback in some corner cases #4884
    • Fix the issue that a removed tombstone store appears again after the PD leader transfer ​​#4941
    • Fix the issue that scheduling cannot start immediately after the PD leader transfer #4769
  • TiDB Dashboard

    • Fix a bug that Top SQL cannot collect the CPU overhead of the SQL statements that were running before the Top SQL feature is enabled #33859
  • TiFlash

    • Fix potential data inconsistency after a lot of INSERT and DELETE operations #4956
  • Tools

    • TiCDC

      • Fix excessive memory usage by optimizing the way DDL schemas are buffered #1386
      • Fix data loss that occurs in special incremental scanning scenarios #5468
    • TiDB Data Migration (DM)

      • Fix the start-time time zone issue and change DM behavior from using the downstream time zone to using the upstream time zone #5471
      • Fix the issue that DM occupies more disk space after the task automatically resumes #3734 #5344
      • Fix the problem that checkpoint flush may cause the data of failed rows to be skipped #5279
      • Fix the issue that in some cases manually executing the filtered DDL in the downstream might cause task resumption failure #5272
      • Fix an issue that the uppercase table cannot be replicated when case-sensitive: true is not set #5255
      • Fix the DM worker panic issue that occurs when the primary key is not first in the index returned by the SHOW CREATE TABLE statement #5159
      • Fix the issue that CPU usage may increase and a large amount of log is printed when GTID is enabled or when the task is automatically resumed #5063
      • Fix the offline option and other usage issues in DM WebUI #4993
      • Fix the issue that incremental tasks fail to start when GTID is empty in the upstream #3731
      • Fix the issue that empty configurations may cause dm-master to panic #3732
    • TiDB Lightning

      • Fix the issue that the precheck does not check local disk resources and cluster availability #34213
      • Fix the issue of incorrect routing for schemas #33381
      • Fix the issue that the PD configuration is not restored correctly when TiDB Lightning panics #31733
      • Fix the issue of Local-backend import failure caused by out-of-bounds data in the auto_increment column #27937
      • Fix the issue of local backend import failure when the auto_random or auto_increment column is null #34208