Skip to content

Commit

Permalink
[Improve][core] Change all common lang packages to common lang3 packa…
Browse files Browse the repository at this point in the history
…ges. (#1346)
  • Loading branch information
chaozwn authored Oct 24, 2022
1 parent 2c88dea commit 57437ce
Show file tree
Hide file tree
Showing 86 changed files with 105 additions and 110 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@
import org.apache.flink.yarn.YarnClientYarnClusterInformationRetriever;
import org.apache.flink.yarn.YarnClusterDescriptor;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.yarn.api.records.ApplicationId;
import org.apache.hadoop.yarn.api.records.ApplicationReport;
import org.apache.hadoop.yarn.api.records.YarnApplicationState;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,6 @@
import org.apache.flink.util.FlinkException;
import org.apache.flink.util.Preconditions;
import org.apache.flink.util.ShutdownHookUtil;
import org.apache.flink.util.StringUtils;
import org.apache.flink.yarn.configuration.YarnConfigOptions;
import org.apache.flink.yarn.configuration.YarnConfigOptionsInternal;
import org.apache.flink.yarn.configuration.YarnDeploymentTarget;
Expand All @@ -69,6 +68,7 @@
import org.apache.flink.yarn.entrypoint.YarnSessionClusterEntrypoint;

import org.apache.commons.collections.ListUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hdfs.DFSConfigKeys;
Expand Down Expand Up @@ -575,7 +575,7 @@ private ClusterClientProvider<ApplicationId> deployInternal(
clusterSpecification
.getConfiguration()
.getString(ConfigConstant.FLINK_PLUGIN_LOAD_MODE_KEY);
if (org.apache.commons.lang3.StringUtils.equalsIgnoreCase(
if (StringUtils.equalsIgnoreCase(
pluginLoadMode, ConstantValue.SHIP_FILE_PLUGIN_LOAD_MODE)) {
List<File> fileList =
jobGraph.getUserArtifacts().entrySet().stream()
Expand Down Expand Up @@ -1074,7 +1074,7 @@ private ApplicationReport startAppMaster(
Path remoteKrb5Path = null;
boolean hasKrb5 = false;
String krb5Config = configuration.get(SecurityOptions.KERBEROS_KRB5_PATH);
if (!StringUtils.isNullOrWhitespaceOnly(krb5Config)) {
if (!StringUtils.isBlank(krb5Config)) {
final File krb5 = new File(krb5Config);
LOG.info(
"Adding KRB5 configuration {} to the AM container local resource bucket",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@
import com.alibaba.fastsql.util.JdbcConstants;
import com.alibaba.otter.canal.protocol.CanalEntry.EventType;
import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.ArrayList;
import java.util.Arrays;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@
import com.google.common.collect.Lists;
import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.collections.MapUtils;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.sql.Connection;
import java.sql.DriverManager;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@
import com.datastax.driver.core.policies.DefaultRetryPolicy;
import com.datastax.driver.core.policies.ExponentialReconnectionPolicy;
import com.google.common.base.Preconditions;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
import com.dtstack.chunjun.connector.jdbc.sink.JdbcOutputFormat;
import com.dtstack.chunjun.connector.jdbc.sink.JdbcOutputFormatBuilder;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

public class ClickhouseOutputFormatBuilder extends JdbcOutputFormatBuilder {

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
import com.dtstack.chunjun.enums.ColumnType;
import com.dtstack.chunjun.enums.Semantic;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.Arrays;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
import org.apache.flink.table.data.RowData;

import com.google.common.collect.Lists;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.elasticsearch.action.search.ClearScrollRequest;
import org.elasticsearch.action.search.ClearScrollResponse;
import org.elasticsearch.action.search.SearchRequest;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@
import org.apache.flink.table.types.logical.RowType;

import com.esotericsoftware.minlog.Log;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.http.util.EntityUtils;
import org.elasticsearch.client.Request;
import org.elasticsearch.client.Response;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@

import com.dtstack.chunjun.connector.ftp.conf.FtpConfig;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.io.IOException;
import java.io.InputStream;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@

import org.apache.flink.table.data.RowData;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.io.BufferedWriter;
import java.io.File;
Expand Down Expand Up @@ -182,7 +182,7 @@ protected void deleteDataFiles(List<String> preCommitFilePathList, String path)
String currentFilePath = "";
try {
for (String filePath : this.preCommitFilePathList) {
if (org.apache.commons.lang3.StringUtils.equals(path, outputFilePath)) {
if (StringUtils.equals(path, outputFilePath)) {
ftpHandler.deleteFile(filePath);
LOG.info("delete file:{}", currentFilePath);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
import com.dtstack.chunjun.sink.format.FileOutputFormatBuilder;
import com.dtstack.chunjun.throwable.ChunJunRuntimeException;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

/**
* The builder of FtpOutputFormat
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@
import org.apache.flink.table.data.RowData;

import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.io.IOException;
import java.util.ArrayList;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
import com.dtstack.chunjun.source.format.BaseRichInputFormatBuilder;
import com.dtstack.chunjun.throwable.ChunJunRuntimeException;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

/** @author jiangbo */
public class FtpInputFormatBuilder extends BaseRichInputFormatBuilder<FtpInputFormat> {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@

package com.dtstack.chunjun.connector.hbase;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

/**
* @company: www.dtstack.com
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
*/
package com.dtstack.chunjun.connector.hbase;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.ArrayList;
import java.util.List;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@
import com.google.common.collect.Lists;
import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.collections.MapUtils;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.io.Serializable;
import java.util.ArrayList;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -129,8 +129,8 @@ public HBaseColumnConverter(HBaseConf hBaseConf, RowType rowType) {
columnNames.add(name);
String[] cfAndQualifier = name.split(":");
if (cfAndQualifier.length == 2
&& org.apache.commons.lang.StringUtils.isNotBlank(cfAndQualifier[0])
&& org.apache.commons.lang.StringUtils.isNotBlank(cfAndQualifier[1])) {
&& StringUtils.isNotBlank(cfAndQualifier[0])
&& StringUtils.isNotBlank(cfAndQualifier[1])) {

byte[][] qualifierKeys = new byte[2][];
qualifierKeys[0] = Bytes.toBytes(cfAndQualifier[0]);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -114,8 +114,8 @@ public HBaseFlatRowConverter(HBaseConf hBaseConf, RowType rowType) {
columnNames.add(name);
String[] cfAndQualifier = name.split(":");
if (cfAndQualifier.length == 2
&& org.apache.commons.lang.StringUtils.isNotBlank(cfAndQualifier[0])
&& org.apache.commons.lang.StringUtils.isNotBlank(cfAndQualifier[1])) {
&& StringUtils.isNotBlank(cfAndQualifier[0])
&& StringUtils.isNotBlank(cfAndQualifier[1])) {

byte[][] qualifierKeys = new byte[2][];
qualifierKeys[0] = Bytes.toBytes(cfAndQualifier[0]);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
import com.dtstack.chunjun.sink.format.BaseRichOutputFormatBuilder;

import com.google.common.base.Preconditions;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.Map;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@
import org.apache.flink.table.types.DataType;
import org.apache.flink.table.types.logical.RowType;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.List;
import java.util.Map;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@

import org.apache.flink.util.Preconditions;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.List;
import java.util.Map;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@
import org.apache.flink.table.types.DataType;
import org.apache.flink.table.types.logical.RowType;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

import java.util.ArrayList;
import java.util.LinkedHashMap;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
*/
package com.dtstack.chunjun.connector.hdfs.enums;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

/**
* Date: 2021/06/09 Company: www.dtstack.com
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
import org.apache.flink.table.data.GenericRowData;
import org.apache.flink.table.data.RowData;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.io.orc.OrcFile;
import org.apache.hadoop.hive.ql.io.orc.OrcInputFormat;
Expand Down Expand Up @@ -65,7 +65,7 @@ public class HdfsOrcInputFormat extends BaseHdfsInputFormat {
public HdfsOrcInputSplit[] createHdfsSplit(int minNumSplits) throws IOException {
super.initHadoopJobConf();
String path;
if (org.apache.commons.lang3.StringUtils.isNotBlank(hdfsConf.getFileName())) {
if (StringUtils.isNotBlank(hdfsConf.getFileName())) {
// 兼容平台逻辑
path = hdfsConf.getPath() + ConstantValue.SINGLE_SLASH_SYMBOL + hdfsConf.getFileName();
} else {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
*/
package com.dtstack.chunjun.connector.hdfs.source;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter;
import org.apache.hadoop.mapred.JobConf;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@
import com.google.common.collect.Lists;
import com.google.common.collect.Maps;
import org.apache.commons.collections.MapUtils;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.security.UserGroupInformation;
import org.slf4j.Logger;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
*/
package com.dtstack.chunjun.connector.hive3.enums;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;

/**
* Date: 2021/06/09 Company: www.dtstack.com
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -68,8 +68,7 @@ protected void openSource() {
}
Object hadoopUser = hdfsConf.getHadoopConfig().get(HADOOP_USER_NAME);
// 如果配置的 hadoop 用户不为空,那么设置配置中的用户。
if (hadoopUser != null
&& org.apache.commons.lang.StringUtils.isNotEmpty(hadoopUser.toString())) {
if (hadoopUser != null && StringUtils.isNotEmpty(hadoopUser.toString())) {
currentUser = hadoopUser.toString();
}
conf =
Expand Down Expand Up @@ -117,8 +116,7 @@ public void getUgi() throws IOException {
openKerberos = Hive3Util.isOpenKerberos(hdfsConf.getHadoopConfig());
String currentUser = UserGroupInformation.getCurrentUser().getUserName();
Object hadoopUser = hdfsConf.getHadoopConfig().get(HADOOP_USER_NAME);
if (hadoopUser != null
&& org.apache.commons.lang.StringUtils.isNotEmpty(hadoopUser.toString())) {
if (hadoopUser != null && StringUtils.isNotEmpty(hadoopUser.toString())) {
currentUser = hadoopUser.toString();
}
if (openKerberos) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
import org.apache.flink.table.data.GenericRowData;
import org.apache.flink.table.data.RowData;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.io.orc.OrcFile;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
*/
package com.dtstack.chunjun.connector.hive3.source;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter;
import org.apache.hadoop.mapred.JobConf;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@

import org.apache.flink.core.io.InputSplit;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -484,7 +484,7 @@ public static boolean isOpenKerberos(Map<String, Object> hadoopConfig) {

public static void setHadoopUserName(Configuration conf) {
String hadoopUserName = conf.get(KEY_HADOOP_USER_NAME);
if (org.apache.commons.lang.StringUtils.isEmpty(hadoopUserName)) {
if (StringUtils.isEmpty(hadoopUserName)) {
return;
}

Expand All @@ -503,7 +503,7 @@ public static void setHadoopUserName(Configuration conf) {

private static FileSystem getFsWithUser(
Map<String, Object> hadoopConfig, String defaultFs, String user) throws Exception {
if (org.apache.commons.lang.StringUtils.isEmpty(user)) {
if (StringUtils.isEmpty(user)) {
return FileSystem.get(getConfiguration(hadoopConfig, defaultFs));
}
UserGroupInformation ugi = UserGroupInformation.createRemoteUser(user);
Expand Down Expand Up @@ -595,7 +595,6 @@ private static Map<String, Object> fillConfig(Map<String, Object> confMap, Strin
}

private static boolean isHaMode(Map<String, Object> confMap) {
return org.apache.commons.lang.StringUtils.isNotEmpty(
MapUtils.getString(confMap, KEY_DFS_NAME_SERVICES));
return StringUtils.isNotEmpty(MapUtils.getString(confMap, KEY_DFS_NAME_SERVICES));
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,8 @@

import com.google.common.collect.Sets;
import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.lang.text.StrBuilder;
import org.apache.commons.lang3.StringUtils;
import org.apache.commons.lang3.text.StrBuilder;

import java.util.ArrayList;
import java.util.Comparator;
Expand Down
Loading

0 comments on commit 57437ce

Please sign in to comment.