diff --git a/docs/content/about.md b/docs/content/about.md index fb4fe11862a..5814abd632c 100644 --- a/docs/content/about.md +++ b/docs/content/about.md @@ -21,17 +21,18 @@ The CDC Connectors for Apache Flink® integrate Debezium as the engin ## Supported Flink Versions The following table shows the version mapping between Flink® CDC Connectors and Flink®: -| Flink® CDC Version | Flink® Version | -|:---:|:---:| -| 1.0.0 | 1.11.* | -| 1.1.0 | 1.11.* | -| 1.2.0 | 1.12.* | -| 1.3.0 | 1.12.* | -| 1.4.0 | 1.13.* | -| 2.0.* | 1.13.* | -| 2.1.* | 1.13.* | -| 2.2.* | 1.13.\*, 1.14.\* | -| 2.3.* | 1.13.\*, 1.14.\*, 1.15.\*, 1.16.0 | +| Flink® CDC Version | Flink® Version | +|:-----------------------------------:|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:| +| 1.0.0 | 1.11.* | +| 1.1.0 | 1.11.* | +| 1.2.0 | 1.12.* | +| 1.3.0 | 1.12.* | +| 1.4.0 | 1.13.* | +| 2.0.* | 1.13.* | +| 2.1.* | 1.13.* | +| 2.2.* | 1.13.\*, 1.14.\* | +| 2.3.* | 1.13.\*, 1.14.\*, 1.15.\*, 1.16.0 | +| 2.4.* | 1.13.\*, 1.14.\*, 1.15.\*, 1.16.\*, 1.17.0 | ## Features diff --git a/docs/content/connectors/mongodb-cdc.md b/docs/content/connectors/mongodb-cdc.md index dd178bc03e6..f1aaa3e63b7 100644 --- a/docs/content/connectors/mongodb-cdc.md +++ b/docs/content/connectors/mongodb-cdc.md @@ -430,9 +430,9 @@ public class MongoDBIncrementalSourceExample { Data Type Mapping ---------------- -[BSON](https://docs.mongodb.com/manual/reference/bson-types/) short for **Binary JSON** is a bin­ary-en­coded seri­al­iz­a­tion of JSON-like format used to store documents and make remote procedure calls in MongoDB. +[BSON](https://docs.mongodb.com/manual/reference/bson-types/) short for **Binary JSON** is a binary-encoded serialization of JSON-like format used to store documents and make remote procedure calls in MongoDB. -[Flink SQL Data Type](https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/dev/table/types/) is similar to the SQL standard’s data type terminology which describes the logical type of a value in the table ecosystem. It can be used to declare input and/or output types of operations. +[Flink SQL Data Type](https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/types/) is similar to the SQL standard’s data type terminology which describes the logical type of a value in the table ecosystem. It can be used to declare input and/or output types of operations. In order to enable Flink SQL to process data from heterogeneous data sources, the data types of heterogeneous data sources need to be uniformly converted to Flink SQL data types. @@ -552,7 +552,7 @@ Reference - [Replica set protocol](https://docs.mongodb.com/manual/reference/replica-configuration/#mongodb-rsconf-rsconf.protocolVersion) - [Connection String Options](https://docs.mongodb.com/manual/reference/connection-string/#std-label-connections-connection-options) - [BSON Types](https://docs.mongodb.com/manual/reference/bson-types/) -- [Flink DataTypes](https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/dev/table/types/) +- [Flink DataTypes](https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/types/) FAQ -------- diff --git a/docs/content/connectors/mysql-cdc(ZH).md b/docs/content/connectors/mysql-cdc(ZH).md index fb0b903b2dd..db5e9f5b88e 100644 --- a/docs/content/connectors/mysql-cdc(ZH).md +++ b/docs/content/connectors/mysql-cdc(ZH).md @@ -688,7 +688,7 @@ $ ./bin/flink run \ --fromSavepoint /tmp/flink-savepoints/savepoint-cca7bc-bb1e257f0dab \ ./FlinkCDCExample.jar ``` -**注意:** 请参考文档 [Restore the job from previous savepoint](https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/deployment/cli/#command-line-interface) 了解更多详细信息。 +**注意:** 请参考文档 [Restore the job from previous savepoint](https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/deployment/cli/#command-line-interface) 了解更多详细信息。 数据类型映射 ---------------- diff --git a/docs/content/connectors/mysql-cdc.md b/docs/content/connectors/mysql-cdc.md index fc33b87ff64..593440921a0 100644 --- a/docs/content/connectors/mysql-cdc.md +++ b/docs/content/connectors/mysql-cdc.md @@ -696,7 +696,7 @@ $ ./bin/flink run \ --fromSavepoint /tmp/flink-savepoints/savepoint-cca7bc-bb1e257f0dab \ ./FlinkCDCExample.jar ``` -**Note:** Please refer the doc [Restore the job from previous savepoint](https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/deployment/cli/#command-line-interface) for more details. +**Note:** Please refer the doc [Restore the job from previous savepoint](https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/deployment/cli/#command-line-interface) for more details. Data Type Mapping ---------------- diff --git a/docs/content/formats/changelog-json.md b/docs/content/formats/changelog-json.md index a199134e051..b31badf25a3 100644 --- a/docs/content/formats/changelog-json.md +++ b/docs/content/formats/changelog-json.md @@ -1,7 +1,7 @@ # Changelog JSON Format **WARNING:** The CDC format `changelog-json` is deprecated since Flink CDC version 2.2. -The CDC format `changelog-json` was introduced at the point that Flink didn't offer any CDC format. Currently, Flink offers several well-maintained CDC formats i.e.[Debezium CDC, MAXWELL CDC, CANAL CDC](https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/connectors/table/formats/overview/), we recommend user to use above CDC formats. +The CDC format `changelog-json` was introduced at the point that Flink didn't offer any CDC format. Currently, Flink offers several well-maintained CDC formats i.e.[Debezium CDC, MAXWELL CDC, CANAL CDC](https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/connectors/table/formats/overview/), we recommend user to use above CDC formats. ### Compatibility Note diff --git a/docs/content/quickstart/db2-tutorial.md b/docs/content/quickstart/db2-tutorial.md index 5be533c2015..ba992fdd59f 100644 --- a/docs/content/quickstart/db2-tutorial.md +++ b/docs/content/quickstart/db2-tutorial.md @@ -61,7 +61,7 @@ docker-compose down *Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. * -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-db2-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-db2-cdc/2.4-SNAPSHOT/flink-sql-connector-db2-cdc-2.4-SNAPSHOT.jar) **3. Launch a Flink cluster and start a Flink SQL CLI** diff --git a/docs/content/quickstart/mongodb-tutorial.md b/docs/content/quickstart/mongodb-tutorial.md index b7628299526..f92d6fdac1c 100644 --- a/docs/content/quickstart/mongodb-tutorial.md +++ b/docs/content/quickstart/mongodb-tutorial.md @@ -109,8 +109,8 @@ db.customers.insertMany([ ```Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. ``` -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) - - [flink-sql-connector-mongodb-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mongodb-cdc/2.4-SNAPSHOT/flink-sql-connector-mongodb-cdc-2.4-SNAPSHOT.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) +- [flink-sql-connector-mongodb-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mongodb-cdc/2.4-SNAPSHOT/flink-sql-connector-mongodb-cdc-2.4-SNAPSHOT.jar) 4. Launch a Flink cluster, then start a Flink SQL CLI and execute following SQL statements inside: diff --git a/docs/content/quickstart/mysql-postgres-tutorial.md b/docs/content/quickstart/mysql-postgres-tutorial.md index 20e5930d2bc..20d6b3e0a8a 100644 --- a/docs/content/quickstart/mysql-postgres-tutorial.md +++ b/docs/content/quickstart/mysql-postgres-tutorial.md @@ -73,11 +73,11 @@ This command automatically starts all the containers defined in the Docker Compo We can also visit [http://localhost:5601/](http://localhost:5601/) to see if Kibana is running normally. ### Preparing Flink and JAR package required -1. Download [Flink 1.16.0](https://archive.apache.org/dist/flink/flink-1.16.0/flink-1.16.0-bin-scala_2.12.tgz) and unzip it to the directory `flink-1.16.0` -2. Download following JAR package required and put them under `flink-1.16.0/lib/`: +1. Download [Flink 1.17.0](https://archive.apache.org/dist/flink/flink-1.17.0/flink-1.17.0-bin-scala_2.12.tgz) and unzip it to the directory `flink-1.17.0` +2. Download following JAR package required and put them under `flink-1.17.0/lib/`: **Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. ** - - [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) + - [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/2.4-SNAPSHOT/flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar) - [flink-sql-connector-postgres-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-postgres-cdc/2.4-SNAPSHOT/flink-sql-connector-postgres-cdc-2.4-SNAPSHOT.jar) @@ -151,7 +151,7 @@ We can also visit [http://localhost:5601/](http://localhost:5601/) to see if Kib 1. Use the following command to change to the Flink directory: ``` - cd flink-1.16.0 + cd flink-1.17.0 ``` 2. Use the following command to start a Flink cluster: @@ -311,7 +311,7 @@ After finishing the tutorial, run the following command to stop all containers i ```shell docker-compose down ``` -Run the following command to stop the Flink cluster in the directory of Flink `flink-1.16.0`: +Run the following command to stop the Flink cluster in the directory of Flink `flink-1.17.0`: ```shell ./bin/stop-cluster.sh ``` diff --git a/docs/content/quickstart/oceanbase-tutorial.md b/docs/content/quickstart/oceanbase-tutorial.md index c02ef34b37d..d1b981034da 100644 --- a/docs/content/quickstart/oceanbase-tutorial.md +++ b/docs/content/quickstart/oceanbase-tutorial.md @@ -111,7 +111,7 @@ VALUES (default, '2020-07-30 10:08:22', 'Jark', 50.50, 102, false), ```Download links are only available for stable releases.``` -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-oceanbase-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-oceanbase-cdc/2.4-SNAPSHOT/flink-sql-connector-oceanbase-cdc-2.4-SNAPSHOT.jar) ### Use Flink DDL to create dynamic table in Flink SQL CLI diff --git a/docs/content/quickstart/oracle-tutorial.md b/docs/content/quickstart/oracle-tutorial.md index 636d3e5ee6d..64261758aeb 100644 --- a/docs/content/quickstart/oracle-tutorial.md +++ b/docs/content/quickstart/oracle-tutorial.md @@ -54,7 +54,7 @@ docker-compose down *Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. * -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-oracle-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-oracle-cdc/2.4-SNAPSHOT/flink-sql-connector-oracle-cdc-2.4-SNAPSHOT.jar) **3. Launch a Flink cluster and start a Flink SQL CLI** diff --git a/docs/content/quickstart/polardbx-tutorial.md b/docs/content/quickstart/polardbx-tutorial.md index 6bf3eab02e1..14508d9636c 100644 --- a/docs/content/quickstart/polardbx-tutorial.md +++ b/docs/content/quickstart/polardbx-tutorial.md @@ -63,12 +63,12 @@ This command automatically starts all the containers defined in the Docker Compo We can also visit [http://localhost:5601/](http://localhost:5601/) to see if Kibana is running normally. ### Preparing Flink and JAR package required -1. Download [Flink 1.16.0](https://archive.apache.org/dist/flink/flink-1.16.0/flink-1.16.0-bin-scala_2.12.tgz) and unzip it to the directory `flink-1.16.0` -2. Download following JAR package required and put them under `flink-1.16.0/lib/`: +1. Download [Flink 1.17.0](https://archive.apache.org/dist/flink/flink-1.17.0/flink-1.17.0-bin-scala_2.12.tgz) and unzip it to the directory `flink-1.17.0` +2. Download following JAR package required and put them under `flink-1.17.0/lib/`: **Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. ** - [flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/2.4-SNAPSHOT/flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar) - - [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) + - [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) ### Preparing data in databases #### Preparing data in PolarDB-X @@ -116,7 +116,7 @@ We can also visit [http://localhost:5601/](http://localhost:5601/) to see if Kib 1. Use the following command to change to the Flink directory: ``` - cd flink-1.16.0 + cd flink-1.17.0 ``` 2. Use the following command to start a Flink cluster: @@ -255,7 +255,7 @@ After finishing the tutorial, run the following command to stop all containers i ```shell docker-compose down ``` -Run the following command to stop the Flink cluster in the directory of Flink `flink-1.16.0`: +Run the following command to stop the Flink cluster in the directory of Flink `flink-1.17.0`: ```shell ./bin/stop-cluster.sh ``` diff --git a/docs/content/quickstart/sqlserver-tutorial.md b/docs/content/quickstart/sqlserver-tutorial.md index 0a9b1d2c706..e17fa353e6b 100644 --- a/docs/content/quickstart/sqlserver-tutorial.md +++ b/docs/content/quickstart/sqlserver-tutorial.md @@ -63,7 +63,7 @@ docker-compose down *Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. * -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-sqlserver-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-sqlserver-cdc/2.4-SNAPSHOT/flink-sql-connector-sqlserver-cdc-2.4-SNAPSHOT.jar) diff --git a/docs/content/quickstart/tidb-tutorial.md b/docs/content/quickstart/tidb-tutorial.md index e9e51aea328..2e70eb610b0 100644 --- a/docs/content/quickstart/tidb-tutorial.md +++ b/docs/content/quickstart/tidb-tutorial.md @@ -116,7 +116,7 @@ docker-compose down *Download links are available only for stable releases, SNAPSHOT dependency need build by yourself. * -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-tidb-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-tidb-cdc/2.4-SNAPSHOT/flink-sql-connector-tidb-cdc-2.4-SNAPSHOT.jar) diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mongodb-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mongodb-tutorial-zh.md" index 6e472fddf57..9172b76900f 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mongodb-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mongodb-tutorial-zh.md" @@ -109,7 +109,7 @@ db.customers.insertMany([ ```下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译``` - - [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) + - [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-mongodb-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mongodb-cdc/2.4-SNAPSHOT/flink-sql-connector-mongodb-cdc-2.4-SNAPSHOT.jar) 4. 然后启动 Flink 集群,再启动 SQL CLI. diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mysql-postgres-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mysql-postgres-tutorial-zh.md" index c153a913273..a895937672e 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mysql-postgres-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/mysql-postgres-tutorial-zh.md" @@ -69,11 +69,11 @@ docker-compose up -d 该命令将以 detached 模式自动启动 Docker Compose 配置中定义的所有容器。你可以通过 docker ps 来观察上述的容器是否正常启动了,也可以通过访问 [http://localhost:5601/](http://localhost:5601/) 来查看 Kibana 是否运行正常。 ### 下载 Flink 和所需要的依赖包 -1. 下载 [Flink 1.16.0](https://archive.apache.org/dist/flink/flink-1.16.0/flink-1.16.0-bin-scala_2.12.tgz) 并将其解压至目录 `flink-1.16.0` -2. 下载下面列出的依赖包,并将它们放到目录 `flink-1.16.0/lib/` 下: +1. 下载 [Flink 1.17.0](https://archive.apache.org/dist/flink/flink-1.17.0/flink-1.17.0-bin-scala_2.12.tgz) 并将其解压至目录 `flink-1.17.0` +2. 下载下面列出的依赖包,并将它们放到目录 `flink-1.17.0/lib/` 下: **下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译** - - [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) + - [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/2.4-SNAPSHOT/flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar) - [flink-sql-connector-postgres-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-postgres-cdc/2.4-SNAPSHOT/flink-sql-connector-postgres-cdc-2.4-SNAPSHOT.jar) @@ -147,7 +147,7 @@ docker-compose up -d 1. 使用下面的命令跳转至 Flink 目录下 ``` - cd flink-1.16.0 + cd flink-1.17.0 ``` 2. 使用下面的命令启动 Flink 集群 @@ -308,7 +308,7 @@ Flink SQL> INSERT INTO enriched_orders ```shell docker-compose down ``` -在 Flink 所在目录 `flink-1.16.0` 下执行如下命令停止 Flink 集群: +在 Flink 所在目录 `flink-1.17.0` 下执行如下命令停止 Flink 集群: ```shell ./bin/stop-cluster.sh ``` diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oceanbase-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oceanbase-tutorial-zh.md" index 358da349873..e2ea68cbb79 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oceanbase-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oceanbase-tutorial-zh.md" @@ -110,7 +110,7 @@ VALUES (default, '2020-07-30 10:08:22', 'Jark', 50.50, 102, false), ```下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译``` -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-oceanbase-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-oceanbase-cdc/2.4-SNAPSHOT/flink-sql-connector-oceanbase-cdc-2.4-SNAPSHOT.jar) ### 在 Flink SQL CLI 中使用 Flink DDL 创建表 diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oracle-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oracle-tutorial-zh.md" index 23ead2aebc8..6af0bb86e61 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oracle-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/oracle-tutorial-zh.md" @@ -54,7 +54,7 @@ docker-compose down *下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译* -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-oracle-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-oracle-cdc/2.4-SNAPSHOT/flink-sql-connector-oracle-cdc-2.4-SNAPSHOT.jar) **3. 然后启动 Flink 集群,再启动 SQL CLI:** diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/polardbx-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/polardbx-tutorial-zh.md" index 0bdc617adf0..b1bc7001f72 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/polardbx-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/polardbx-tutorial-zh.md" @@ -105,12 +105,12 @@ VALUES (default, '2020-07-30 10:08:22', 'Jark', 50.50, 102, false), ``` ### 下载 Flink 和所需要的依赖包 -1. 下载 [Flink 1.16.0](https://archive.apache.org/dist/flink/flink-1.16.0/flink-1.16.0-bin-scala_2.12.tgz) 并将其解压至目录 `flink-1.16.0` -2. 下载下面列出的依赖包,并将它们放到目录 `flink-1.16.0/lib/` 下 +1. 下载 [Flink 1.17.0](https://archive.apache.org/dist/flink/flink-1.17.0/flink-1.17.0-bin-scala_2.12.tgz) 并将其解压至目录 `flink-1.17.0` +2. 下载下面列出的依赖包,并将它们放到目录 `flink-1.17.0/lib/` 下 ```下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译``` - 用于订阅PolarDB-X Binlog: [flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/2.4-SNAPSHOT/flink-sql-connector-mysql-cdc-2.4-SNAPSHOT.jar) -- 用于写入Elasticsearch: [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- 用于写入Elasticsearch: [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) 3. 启动flink服务: ```shell ./bin/start-cluster.sh diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/sqlserver-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/sqlserver-tutorial-zh.md" index f2b4d2afb66..85226a716f8 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/sqlserver-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/sqlserver-tutorial-zh.md" @@ -63,7 +63,7 @@ docker-compose down ```下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译``` -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-sqlserver-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-sqlserver-cdc/2.4-SNAPSHOT/flink-sql-connector-sqlserver-cdc-2.4-SNAPSHOT.jar) diff --git "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/tidb-tutorial-zh.md" "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/tidb-tutorial-zh.md" index a949d1e4fa6..70f2ff665f4 100644 --- "a/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/tidb-tutorial-zh.md" +++ "b/docs/content/\345\277\253\351\200\237\344\270\212\346\211\213/tidb-tutorial-zh.md" @@ -116,7 +116,7 @@ docker-compose down ```下载链接只对已发布的版本有效, SNAPSHOT 版本需要本地编译``` -- [flink-sql-connector-elasticsearch7-1.16.0.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/1.16.0/flink-sql-connector-elasticsearch7-1.16.0.jar) +- [flink-sql-connector-elasticsearch7-3.0.1-1.17.jar](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7/3.0.1-1.17/flink-sql-connector-elasticsearch7-3.0.1-1.17.jar) - [flink-sql-connector-tidb-cdc-2.4-SNAPSHOT.jar](https://repo1.maven.org/maven2/com/ververica/flink-sql-connector-tidb-cdc/2.4-SNAPSHOT/flink-sql-connector-tidb-cdc-2.4-SNAPSHOT.jar) diff --git a/flink-cdc-e2e-tests/pom.xml b/flink-cdc-e2e-tests/pom.xml index 56a93727c3d..5a63d3b43d8 100644 --- a/flink-cdc-e2e-tests/pom.xml +++ b/flink-cdc-e2e-tests/pom.xml @@ -32,6 +32,8 @@ under the License. 1.14.4 1.15.2 1.16.0 + 1.17.0 + 3.1.0-1.17 8.0.27 @@ -249,6 +251,16 @@ under the License. + + org.apache.flink + flink-connector-jdbc + ${jdbc.version-1.17} + jdbc-connector_${flink-1.17}.jar + jar + ${project.build.directory}/dependencies + + + com.ververica flink-sql-connector-mysql-cdc diff --git a/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/MySqlE2eITCase.java b/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/MySqlE2eITCase.java index 5e8fe3b9327..f35cc45dc3f 100644 --- a/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/MySqlE2eITCase.java +++ b/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/MySqlE2eITCase.java @@ -43,6 +43,7 @@ public class MySqlE2eITCase extends FlinkContainerTestEnvironment { public void testMySqlCDC() throws Exception { List sqlLines = Arrays.asList( + "SET 'execution.checkpointing.interval' = '3s';", "CREATE TABLE products_source (", " `id` INT NOT NULL,", " name STRING,", diff --git a/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/utils/FlinkContainerTestEnvironment.java b/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/utils/FlinkContainerTestEnvironment.java index b56d92a477d..52ae9e04f99 100644 --- a/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/utils/FlinkContainerTestEnvironment.java +++ b/flink-cdc-e2e-tests/src/test/java/com/ververica/cdc/connectors/tests/utils/FlinkContainerTestEnvironment.java @@ -121,11 +121,11 @@ public abstract class FlinkContainerTestEnvironment extends TestLogger { @Parameterized.Parameters(name = "flinkVersion: {0}") public static List getFlinkVersion() { - return Arrays.asList("1.13.6", "1.14.4", "1.15.2", "1.16.0"); + return Arrays.asList("1.13.6", "1.14.4", "1.15.2", "1.16.0", "1.17.0"); } private static final List FLINK_VERSION_WITH_SCALA_212 = - Arrays.asList("1.15.2", "1.16.0"); + Arrays.asList("1.15.2", "1.16.0", "1.17.0"); @Before public void before() { diff --git a/flink-connector-postgres-cdc/src/test/java/com/ververica/cdc/connectors/postgres/table/PostgreSQLConnectorITCase.java b/flink-connector-postgres-cdc/src/test/java/com/ververica/cdc/connectors/postgres/table/PostgreSQLConnectorITCase.java index 3849b028899..6d47ba77f6b 100644 --- a/flink-connector-postgres-cdc/src/test/java/com/ververica/cdc/connectors/postgres/table/PostgreSQLConnectorITCase.java +++ b/flink-connector-postgres-cdc/src/test/java/com/ververica/cdc/connectors/postgres/table/PostgreSQLConnectorITCase.java @@ -451,10 +451,11 @@ public void testUpsertMode() throws Exception { String sourceDDL = String.format( "CREATE TABLE debezium_source (" - + " id INT NOT NULL PRIMARY KEY," + + " id INT NOT NULL," + " name STRING," + " description STRING," - + " weight DECIMAL(10,3)" + + " weight DECIMAL(10,3)," + + " PRIMARY KEY (id) NOT ENFORCED" + ") WITH (" + " 'connector' = 'postgres-cdc'," + " 'hostname' = '%s'," diff --git a/pom.xml b/pom.xml index 91d144e0ba7..4cefd70f1a3 100644 --- a/pom.xml +++ b/pom.xml @@ -73,7 +73,7 @@ under the License. - 1.16.0 + 1.17.0 1.6.4.Final 3.2.0 2.2.0 @@ -137,11 +137,11 @@ under the License. ${slf4j.version} - + org.apache.flink flink-shaded-guava - 30.1.1-jre-16.0 + 30.1.1-jre-16.1 diff --git a/tools/azure-pipelines/jobs-template.yml b/tools/azure-pipelines/jobs-template.yml index b378e846b18..ba858d21364 100644 --- a/tools/azure-pipelines/jobs-template.yml +++ b/tools/azure-pipelines/jobs-template.yml @@ -66,7 +66,7 @@ jobs: dependsOn: compile_${{parameters.stage_name}} condition: and(succeeded(), not(eq(variables['MODE'], 'e2e'))) pool: ${{parameters.test_pool_definition}} - timeoutInMinutes: 40 + timeoutInMinutes: 70 cancelTimeoutInMinutes: 1 workspace: clean: all