Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 4 additions & 4 deletions common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
Original file line number Diff line number Diff line change
Expand Up @@ -1523,7 +1523,7 @@ public static enum ConfVars {
"This limits the number of partitions that can be requested from the metastore for a given table.\n" +
"The default value \"-1\" means no limit."),

NEWTABLEDEFAULTPARA("hive.table.parameters.default", "",
NEW_TABLE_DEFAULT_PARA("hive.table.parameters.default", "",
"Default property values for newly created tables"),
/**
* @deprecated With HIVE-25813 table properties of source tables will not be copied over to dest table.
Expand Down Expand Up @@ -2813,11 +2813,11 @@ public static enum ConfVars {
"Set this to true to use approximation based logic to adjust ndv after join."),
HIVE_STATS_NUM_NULLS_ESTIMATE_PERC("hive.stats.num.nulls.estimate.percent", (float)5,
"This many percentage of rows will be estimated as number of nulls in absence of statistics."),
HIVESTATSAUTOGATHER("hive.stats.autogather", true,
HIVE_STATS_AUTOGATHER("hive.stats.autogather", true,
"A flag to gather statistics (only basic) automatically during the INSERT OVERWRITE command."),
HIVESTATSCOLAUTOGATHER("hive.stats.column.autogather", true,
HIVE_STATS_COL_AUTOGATHER("hive.stats.column.autogather", true,
"A flag to gather column statistics automatically."),
HIVESTATSDBCLASS("hive.stats.dbclass", "fs", new PatternSet("custom", "fs"),
HIVE_STATS_DBCLASS("hive.stats.dbclass", "fs", new PatternSet("custom", "fs"),
"The storage that stores temporary Hive statistics. In filesystem based statistics collection ('fs'), \n" +
"each task writes statistics it has collected in a file on the filesystem, which will be aggregated \n" +
"after the job has finished. Supported values are fs (filesystem) and custom as defined in StatsSetupConst.java."), // StatsSetupConst.StatDB
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@ public void testAnalyzeTableComputeStatisticsEmptyTable() throws IOException, TE
public void testStatsWithInsert() {
TableIdentifier identifier = TableIdentifier.of("default", "customers");

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA,
PartitionSpec.unpartitioned(), fileFormat, ImmutableList.of());

Expand All @@ -107,7 +107,7 @@ public void testStatsWithInsert() {
public void testStatsWithInsertOverwrite() {
TableIdentifier identifier = TableIdentifier.of("default", "customers");

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA,
PartitionSpec.unpartitioned(), fileFormat, ImmutableList.of());

Expand All @@ -125,7 +125,7 @@ public void testStatsWithPartitionedInsert() {
PartitionSpec spec = PartitionSpec.builderFor(HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA)
.identity("last_name").build();

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA, spec,
fileFormat, ImmutableList.of());

Expand All @@ -150,7 +150,7 @@ public void testStatsWithCTAS() {
shell.executeStatement(testTables.getInsertQuery(
HiveIcebergStorageHandlerTestUtils.CUSTOMER_RECORDS, TableIdentifier.of("default", "source"), false));

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
shell.executeStatement(String.format(
"CREATE TABLE target STORED BY ICEBERG %s %s AS SELECT * FROM source",
testTables.locationForCreateTableSQL(TableIdentifier.of("default", "target")),
Expand All @@ -169,7 +169,7 @@ public void testStatsWithPartitionedCTAS() {
shell.executeStatement(testTables.getInsertQuery(
HiveIcebergStorageHandlerTestUtils.CUSTOMER_RECORDS, TableIdentifier.of("default", "source"), false));

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
shell.executeStatement(String.format(
"CREATE TABLE target PARTITIONED BY (dept, name) STORED BY ICEBERG %s AS SELECT * FROM source s",
testTables.propertiesForCreateTableSQL(
Expand All @@ -189,7 +189,7 @@ public void testStatsRemoved() throws IOException {

TableIdentifier identifier = TableIdentifier.of("default", "customers");

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA,
PartitionSpec.unpartitioned(), fileFormat, ImmutableList.of());

Expand All @@ -215,7 +215,7 @@ public void testStatsRemoved() throws IOException {
public void testColumnStatsAccurate() throws Exception {
TableIdentifier identifier = TableIdentifier.of("default", "customers");

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA,
PartitionSpec.unpartitioned(), fileFormat, ImmutableList.of());

Expand All @@ -236,7 +236,7 @@ public void testColumnStatsAccurate() throws Exception {
public void testMergeStatsWithInsert() {
TableIdentifier identifier = TableIdentifier.of("default", "customers");

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA,
PartitionSpec.unpartitioned(), fileFormat, ImmutableList.of());

Expand Down Expand Up @@ -267,7 +267,7 @@ public void testMergeStatsWithInsert() {
public void testIcebergColStatsPath() throws IOException {
TableIdentifier identifier = TableIdentifier.of("default", "customers");

shell.setHiveSessionValue(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname, true);
shell.setHiveSessionValue(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname, true);
Table table = testTables.createTable(shell, identifier.name(), HiveIcebergStorageHandlerTestUtils.CUSTOMER_SCHEMA,
PartitionSpec.unpartitioned(), fileFormat, ImmutableList.of());

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -202,7 +202,7 @@ private HiveConf initializeConf() {

// Speed up test execution
hiveConf.setLongVar(HiveConf.ConfVars.HIVE_COUNTERS_PULL_INTERVAL, 1L);
hiveConf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);

// Resource configuration
hiveConf.setInt("mapreduce.map.memory.mb", 1024);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -146,7 +146,7 @@ public static void setUpBeforeClass() throws Exception {
hiveConf.setVar(ConfVars.SEMANTIC_ANALYZER_HOOK,
SemanticAnalysisHook.class.getName());
hiveConf.setBoolVar(ConfVars.HIVE_SUPPORT_CONCURRENCY, false);
hiveConf.setBoolVar(ConfVars.HIVESTATSCOLAUTOGATHER, false);
hiveConf.setBoolVar(ConfVars.HIVE_STATS_COL_AUTOGATHER, false);

hiveServer2 = new HiveServer2();
hiveServer2.init(hiveConf);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,6 @@

import java.io.ByteArrayOutputStream;
import java.io.File;
import java.io.IOException;
import java.io.OutputStream;
import java.io.PrintStream;
import java.nio.file.Files;
Expand Down Expand Up @@ -117,7 +116,7 @@ public void setUp() throws Exception {
"org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdHiveAuthorizerFactory");
hiveConf.setBoolVar(HiveConf.ConfVars.MERGE_CARDINALITY_VIOLATION_CHECK, true);
HiveConf.setBoolVar(hiveConf, HiveConf.ConfVars.SPLIT_UPDATE, true);
hiveConf.setBoolVar(HiveConf.ConfVars.HIVESTATSCOLAUTOGATHER, false);
hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_COL_AUTOGATHER, false);
hiveConf.setBoolean("mapred.input.dir.recursive", true);
TestTxnDbUtil.setConfValues(hiveConf);
txnHandler = TxnUtils.getTxnStore(hiveConf);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -221,7 +221,7 @@ static void internalBeforeClassSetup(Map<String, String> additionalProperties)
"org.apache.hadoop.hive.metastore.InjectableBehaviourObjectStore");
hconf.set(HiveConf.ConfVars.HIVE_METASTORE_WAREHOUSE_EXTERNAL.varname, "/tmp/warehouse/external");
hconf.setBoolVar(HiveConf.ConfVars.HIVE_OPTIMIZE_METADATA_QUERIES, true);
hconf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, true);
hconf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, true);
hconf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_RELIABLE, true);
hconf.setBoolVar(HiveConf.ConfVars.REPL_RUN_DATA_COPY_TASKS_ON_TARGET, false);
hconf.setBoolVar(HiveConf.ConfVars.REPL_BATCH_INCREMENTAL_EVENTS, false);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,7 @@ static void internalBeforeClassSetup(Map<String, String> primaryOverrides,
// Run with autogather false on primary if requested
Map<String, String> sourceOverrides = new HashMap<>();
hasAutogather = autogather;
additionalOverrides.put(HiveConf.ConfVars.HIVESTATSAUTOGATHER.varname,
additionalOverrides.put(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER.varname,
autogather ? "true" : "false");
sourceOverrides.putAll(additionalOverrides);
sourceOverrides.putAll(primaryOverrides);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ public void setup() throws Exception {
hiveConf.setVar(HiveConf.ConfVars.POST_EXEC_HOOKS, "");
hiveConf.setVar(HiveConf.ConfVars.METASTORE_WAREHOUSE, TEST_WAREHOUSE_DIR);
hiveConf.setVar(HiveConf.ConfVars.HIVE_INPUT_FORMAT, HiveInputFormat.class.getName());
hiveConf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);
hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_OPTIMIZE_METADATA_QUERIES, false);
MetastoreConf.setBoolVar(hiveConf, MetastoreConf.ConfVars.COMPACTOR_INITIATOR_ON, true);
MetastoreConf.setBoolVar(hiveConf, MetastoreConf.ConfVars.COMPACTOR_CLEANER_ON, true);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -105,7 +105,7 @@ public void testRebalanceCompactionWithParallelDeleteAsSecondPessimisticLock() t
private void testRebalanceCompactionWithParallelDeleteAsSecond(boolean optimisticLock) throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, true);
conf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.TXN_WRITE_X_LOCK, optimisticLock);

//set grouping size to have 3 buckets, and re-create driver with the new config
Expand Down Expand Up @@ -204,7 +204,7 @@ private void testRebalanceCompactionWithParallelDeleteAsSecond(boolean optimisti
public void testRebalanceCompactionOfNotPartitionedImplicitlyBucketedTableWithOrder() throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, true);
conf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);

//set grouping size to have 3 buckets, and re-create driver with the new config
conf.set("tez.grouping.min-size", "400");
Expand Down Expand Up @@ -260,7 +260,7 @@ public void testRebalanceCompactionOfNotPartitionedImplicitlyBucketedTableWithOr
public void testRebalanceCompactionOfNotPartitionedImplicitlyBucketedTable() throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, true);
conf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);

//set grouping size to have 3 buckets, and re-create driver with the new config
conf.set("tez.grouping.min-size", "400");
Expand Down Expand Up @@ -314,7 +314,7 @@ public void testRebalanceCompactionOfNotPartitionedImplicitlyBucketedTable() thr
public void testRebalanceCompactionOfPartitionedImplicitlyBucketedTable() throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, true);
conf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);

//set grouping size to have 3 buckets, and re-create driver with the new config
conf.set("tez.grouping.min-size", "1");
Expand Down Expand Up @@ -423,7 +423,7 @@ public void testRebalanceCompactionOfPartitionedImplicitlyBucketedTable() throws
public void testRebalanceCompactionOfNotPartitionedExplicitlyBucketedTable() throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, true);
conf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);

final String tableName = "rebalance_test";
executeStatementOnDriver("drop table if exists " + tableName, driver);
Expand All @@ -445,7 +445,7 @@ public void testRebalanceCompactionOfNotPartitionedExplicitlyBucketedTable() thr
public void testRebalanceCompactionNotPartitionedExplicitBucketNumbers() throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, true);
conf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);

//set grouping size to have 3 buckets, and re-create driver with the new config
conf.set("tez.grouping.min-size", "400");
Expand Down Expand Up @@ -3098,7 +3098,7 @@ public void testStatsAfterCompactionPartTblForQueryBasedCompaction() throws Exce
public void testStatsAfterCompactionPartTbl(boolean isQueryBased, boolean isAutoGatherStats,
CompactionType compactionType) throws Exception {
conf.setBoolVar(HiveConf.ConfVars.COMPACTOR_CRUD_QUERY_BASED, isQueryBased);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, isAutoGatherStats);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, isAutoGatherStats);
String dbName = "default";
String tblName = "minor_compaction_test";
IMetaStoreClient msClient = new HiveMetaStoreClient(conf);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ public void testCompactorGatherStats() throws Exception{
conf.setBoolVar(HiveConf.ConfVars.HIVE_WRITE_ACID_VERSION_FILE, true);
conf.setVar(HiveConf.ConfVars.COMPACTOR_JOB_QUEUE, CUSTOM_COMPACTION_QUEUE);
conf.setBoolVar(HiveConf.ConfVars.HIVE_MR_COMPACTOR_GATHER_STATS, true);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER, false);
conf.setVar(HiveConf.ConfVars.HIVE_PROTO_EVENTS_BASE_PATH, tmpFolder);

String dbName = "default";
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -215,7 +215,7 @@ private static void startMiniHS2(HiveConf conf) throws Exception {
private static void startMiniHS2(HiveConf conf, boolean httpMode) throws Exception {
conf.setBoolVar(ConfVars.HIVE_SUPPORT_CONCURRENCY, false);
conf.setBoolVar(ConfVars.HIVE_SERVER2_LOGGING_OPERATION_ENABLED, false);
conf.setBoolVar(ConfVars.HIVESTATSCOLAUTOGATHER, false);
conf.setBoolVar(ConfVars.HIVE_STATS_COL_AUTOGATHER, false);
// store post-exec hooks calls so we can look at them later
conf.setVar(ConfVars.POST_EXEC_HOOKS, ReadableHook.class.getName() + "," +
LineageLogger.class.getName());
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -116,7 +116,7 @@ public static void setupHS2() throws Exception {
HiveConf conf = new HiveConf();
conf.setBoolVar(ConfVars.HIVE_SUPPORT_CONCURRENCY, false);
conf.setBoolVar(ConfVars.HIVE_SERVER2_LOGGING_OPERATION_ENABLED, false);
conf.setBoolVar(ConfVars.HIVESTATSCOLAUTOGATHER, false);
conf.setBoolVar(ConfVars.HIVE_STATS_COL_AUTOGATHER, false);
conf.setVar(ConfVars.HIVE_SERVER2_AUTHENTICATION, "JWT");
// the content of the URL below is the same as jwtVerificationJWKSFile
conf.setVar(ConfVars.HIVE_SERVER2_AUTHENTICATION_JWT_JWKS_URL, "http://localhost:" + MOCK_JWKS_SERVER_PORT +
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -33,9 +33,7 @@
import java.io.File;
import java.io.IOException;
import java.net.InetAddress;
import java.net.InetSocketAddress;
import java.net.ServerSocket;
import java.net.Socket;
import java.nio.charset.StandardCharsets;
import java.sql.ResultSet;
import java.sql.SQLException;
Expand All @@ -53,16 +51,13 @@
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
import org.apache.hadoop.hive.metastore.MetaStoreTestUtils;
import org.apache.hadoop.hive.ql.metadata.TestHive;
import org.apache.hive.jdbc.HiveConnection;
import org.apache.hive.jdbc.Utils.JdbcConnectionParams;
import org.apache.hive.jdbc.miniHS2.MiniHS2;
import org.apache.hive.jdbc.saml.HiveJdbcBrowserClient;
import org.apache.hive.jdbc.saml.IJdbcBrowserClient;
import org.apache.hive.jdbc.saml.IJdbcBrowserClient.HiveJdbcBrowserException;
import org.apache.hive.jdbc.saml.IJdbcBrowserClientFactory;
import org.apache.hive.jdbc.saml.SimpleSAMLPhpTestBrowserClient;
import org.apache.thrift.TException;
import org.junit.After;
import org.junit.AfterClass;
import org.junit.BeforeClass;
Expand Down Expand Up @@ -108,7 +103,7 @@ public static void setupHS2() throws Exception {
HiveConf conf = new HiveConf();
conf.setBoolVar(ConfVars.HIVE_SUPPORT_CONCURRENCY, false);
conf.setBoolVar(ConfVars.HIVE_SERVER2_LOGGING_OPERATION_ENABLED, false);
conf.setBoolVar(ConfVars.HIVESTATSCOLAUTOGATHER, false);
conf.setBoolVar(ConfVars.HIVE_STATS_COL_AUTOGATHER, false);
conf.setVar(ConfVars.HIVE_SERVER2_AUTHENTICATION, "SAML");
conf.setVar(ConfVars.HIVE_SERVER2_SAML_IDP_METADATA,
idpMetadataFile.getAbsolutePath());
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@ public static void setupBeforeClass() throws Exception {
HiveConf conf = new HiveConf();
conf.setBoolVar(HiveConf.ConfVars.HIVE_SUPPORT_CONCURRENCY, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_SERVER2_LOGGING_OPERATION_ENABLED, false);
conf.setBoolVar(HiveConf.ConfVars.HIVESTATSCOLAUTOGATHER, false);
conf.setBoolVar(HiveConf.ConfVars.HIVE_STATS_COL_AUTOGATHER, false);
conf.setTimeVar(HiveConf.ConfVars.HIVE_SERVER2_GRACEFUL_STOP_TIMEOUT, 60, TimeUnit.SECONDS);
MiniHS2.Builder builder = new MiniHS2.Builder().withConf(conf).cleanupLocalDirOnStartup(false);
miniHS2 = builder.build();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -929,7 +929,7 @@ public Table toTable(HiveConf conf) throws HiveException {
StatsSetupConst.setStatsStateForCreateTable(tbl.getTTable().getParameters(), null,
StatsSetupConst.FALSE);
if (!this.isCTAS && !tbl.isPartitioned() && !tbl.isTemporary() &&
conf.getBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER)) {
conf.getBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER)) {
// Put the flag into the dictionary in order not to pollute the table,
// ObjectDictionary is meant to convey repeatitive messages.
ObjectDictionary dictionary = tbl.getTTable().isSetDictionary() ?
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ public int execute() throws HiveException {
}

if (desc.getLocation() == null && !tbl.isPartitioned() &&
context.getConf().getBoolVar(HiveConf.ConfVars.HIVESTATSAUTOGATHER)) {
context.getConf().getBoolVar(HiveConf.ConfVars.HIVE_STATS_AUTOGATHER)) {
StatsSetupConst.setStatsStateForCreateTable(tbl.getTTable().getParameters(),
MetaStoreUtils.getColumnNames(tbl.getCols()), StatsSetupConst.TRUE);
}
Expand Down
Loading