diff --git accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/AccumuloHiveRow.java accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/AccumuloHiveRow.java index 144afe3c7a..b18a80a70a 100644 --- accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/AccumuloHiveRow.java +++ accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/AccumuloHiveRow.java @@ -25,7 +25,7 @@ import java.util.Collections; import java.util.List; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Writable; diff --git accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java index b06b44aa5c..b64dac994e 100644 --- accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java +++ accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java @@ -20,7 +20,7 @@ import java.util.Collections; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.accumulo.AccumuloHiveConstants; import org.apache.hadoop.hive.accumulo.serde.TooManyAccumuloColumnsException; import org.apache.hadoop.hive.serde.serdeConstants; diff --git accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloMapColumnMapping.java accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloMapColumnMapping.java index b2082e8c3c..450ecda5cb 100644 --- accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloMapColumnMapping.java +++ accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloMapColumnMapping.java @@ -16,7 +16,7 @@ */ package org.apache.hadoop.hive.accumulo.columns; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.accumulo.AccumuloHiveConstants; import com.google.common.base.Preconditions; diff --git accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/columns/TestColumnMapper.java accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/columns/TestColumnMapper.java index e5f1b97bbd..dc449e4914 100644 --- accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/columns/TestColumnMapper.java +++ accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/columns/TestColumnMapper.java @@ -20,7 +20,7 @@ import java.util.Iterator; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.accumulo.AccumuloHiveConstants; import org.apache.hadoop.hive.accumulo.serde.TooManyAccumuloColumnsException; import org.apache.hadoop.hive.serde.serdeConstants; diff --git accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java index ed28e18140..abc0ee6024 100644 --- accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java +++ accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java @@ -19,7 +19,7 @@ import java.util.Arrays; import java.util.Properties; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.serde2.lazy.objectinspector.LazySimpleStructObjectInspector; import org.slf4j.Logger; diff --git beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java index d425679b8b..968bdf718e 100644 --- beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java +++ beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java @@ -23,8 +23,8 @@ package org.apache.hive.beeline; import org.apache.commons.io.output.StringBuilderWriter; -import org.apache.commons.lang.BooleanUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.BooleanUtils; +import org.apache.commons.lang3.StringUtils; import org.supercsv.encoder.CsvEncoder; import org.supercsv.encoder.DefaultCsvEncoder; import org.supercsv.encoder.SelectiveCsvEncoder; diff --git cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java index 65062aeed4..cdd08ce7c9 100644 --- cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java +++ cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java @@ -37,7 +37,7 @@ import java.util.Map; import java.util.Set; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git common/src/java/org/apache/hadoop/hive/common/FileUtils.java common/src/java/org/apache/hadoop/hive/common/FileUtils.java index 61aca56bac..cef8fde594 100644 --- common/src/java/org/apache/hadoop/hive/common/FileUtils.java +++ common/src/java/org/apache/hadoop/hive/common/FileUtils.java @@ -1034,7 +1034,7 @@ public static URI getURI(String path) throws URISyntaxException { * @return the list of the file names in the format of URI formats. */ public static Set getJarFilesByPath(String pathString, Configuration conf) { - if (org.apache.commons.lang.StringUtils.isBlank(pathString)) { + if (org.apache.commons.lang3.StringUtils.isBlank(pathString)) { return Collections.emptySet(); } Set result = new HashSet<>(); diff --git common/src/java/org/apache/hadoop/hive/common/LogUtils.java common/src/java/org/apache/hadoop/hive/common/LogUtils.java index 874a3e1274..d409a2133a 100644 --- common/src/java/org/apache/hadoop/hive/common/LogUtils.java +++ common/src/java/org/apache/hadoop/hive/common/LogUtils.java @@ -219,10 +219,13 @@ public static String maskIfPassword(String key, String value) { * Register logging context so that log system can print QueryId, SessionId, etc for each message */ public static void registerLoggingContext(Configuration conf) { - MDC.put(SESSIONID_LOG_KEY, HiveConf.getVar(conf, HiveConf.ConfVars.HIVESESSIONID)); - MDC.put(QUERYID_LOG_KEY, HiveConf.getVar(conf, HiveConf.ConfVars.HIVEQUERYID)); if (HiveConf.getBoolVar(conf, HiveConf.ConfVars.HIVE_SERVER2_LOGGING_OPERATION_ENABLED)) { + MDC.put(SESSIONID_LOG_KEY, HiveConf.getVar(conf, HiveConf.ConfVars.HIVESESSIONID)); + MDC.put(QUERYID_LOG_KEY, HiveConf.getVar(conf, HiveConf.ConfVars.HIVEQUERYID)); MDC.put(OPERATIONLOG_LEVEL_KEY, HiveConf.getVar(conf, HiveConf.ConfVars.HIVE_SERVER2_LOGGING_OPERATION_LEVEL)); + l4j.info("Thread context registration is done."); + } else { + l4j.info("Thread context registration is skipped."); } } @@ -230,7 +233,11 @@ public static void registerLoggingContext(Configuration conf) { * Unregister logging context */ public static void unregisterLoggingContext() { - MDC.clear(); + // Remove the keys added, don't use clear, as it may clear all other things which are not intended to be removed. + MDC.remove(SESSIONID_LOG_KEY); + MDC.remove(QUERYID_LOG_KEY); + MDC.remove(OPERATIONLOG_LEVEL_KEY); + l4j.info("Unregistered logging context."); } /** diff --git common/src/java/org/apache/hadoop/hive/common/cli/HiveFileProcessor.java common/src/java/org/apache/hadoop/hive/common/cli/HiveFileProcessor.java index 62d85605bd..1a6f37188f 100644 --- common/src/java/org/apache/hadoop/hive/common/cli/HiveFileProcessor.java +++ common/src/java/org/apache/hadoop/hive/common/cli/HiveFileProcessor.java @@ -21,7 +21,7 @@ import java.io.BufferedReader; import java.io.IOException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.io.IOUtils; /** diff --git common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java index 3476012523..1a0d7e6a27 100644 --- common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java +++ common/src/java/org/apache/hadoop/hive/common/format/datetime/HiveSqlDateTimeFormatter.java @@ -23,8 +23,7 @@ import com.google.common.base.Preconditions; import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableMap; -import org.apache.commons.lang.StringUtils; -import org.apache.commons.lang.WordUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.common.type.Timestamp; @@ -1414,6 +1413,6 @@ public String getPattern() { } private static String capitalize(String substring) { - return WordUtils.capitalize(substring.toLowerCase()); + return StringUtils.capitalize(substring.toLowerCase()); } } diff --git common/src/java/org/apache/hadoop/hive/common/log/InPlaceUpdate.java common/src/java/org/apache/hadoop/hive/common/log/InPlaceUpdate.java index 37cc12d303..767edcf715 100644 --- common/src/java/org/apache/hadoop/hive/common/log/InPlaceUpdate.java +++ common/src/java/org/apache/hadoop/hive/common/log/InPlaceUpdate.java @@ -20,7 +20,7 @@ import com.google.common.base.Function; import com.google.common.collect.Lists; import jline.TerminalFactory; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.fusesource.jansi.Ansi; diff --git common/src/java/org/apache/hadoop/hive/common/type/HiveBaseChar.java common/src/java/org/apache/hadoop/hive/common/type/HiveBaseChar.java index 2bb2ca24ca..d9ed3e0777 100644 --- common/src/java/org/apache/hadoop/hive/common/type/HiveBaseChar.java +++ common/src/java/org/apache/hadoop/hive/common/type/HiveBaseChar.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.common.type; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; public abstract class HiveBaseChar { protected String value; diff --git common/src/java/org/apache/hadoop/hive/common/type/HiveChar.java common/src/java/org/apache/hadoop/hive/common/type/HiveChar.java index f0b28c720d..f4600a353b 100644 --- common/src/java/org/apache/hadoop/hive/common/type/HiveChar.java +++ common/src/java/org/apache/hadoop/hive/common/type/HiveChar.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.common.type; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; /** * HiveChar. diff --git common/src/java/org/apache/hadoop/hive/conf/HiveConf.java common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index fc53ed397e..7084b0b407 100644 --- common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -22,7 +22,7 @@ import com.google.common.collect.ImmutableSet; import com.google.common.collect.Iterables; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.FileUtils; import org.apache.hadoop.hive.common.ZooKeeperHiveHelper; @@ -2644,6 +2644,8 @@ private static void populateLlapDaemonVarsSet(Set llapDaemonVarsSetLocal "In nonstrict mode, for non-ACID resources, INSERT will only acquire shared lock, which\n" + "allows two concurrent writes to the same partition but still lets lock manager prevent\n" + "DROP TABLE etc. when the table is being written to"), + HIVE_TXN_READ_LOCKS("hive.txn.read.locks", true, + "flag to turn off the strict read lock when set to false"), TXN_OVERWRITE_X_LOCK("hive.txn.xlock.iow", true, "Ensures commands with OVERWRITE (such as INSERT OVERWRITE) acquire Exclusive locks for\n" + "transactional tables. This ensures that inserts (w/o overwrite) running concurrently\n" + diff --git common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java index ebe64234c0..a28580cba1 100644 --- common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java +++ common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.conf; import com.google.common.collect.Iterables; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.classification.InterfaceAudience.Private; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; diff --git common/src/java/org/apache/hive/common/util/HiveStringUtils.java common/src/java/org/apache/hive/common/util/HiveStringUtils.java index 196b9c457b..22948e38de 100644 --- common/src/java/org/apache/hive/common/util/HiveStringUtils.java +++ common/src/java/org/apache/hive/common/util/HiveStringUtils.java @@ -40,7 +40,7 @@ import java.util.regex.Pattern; import com.google.common.base.Splitter; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.text.translate.CharSequenceTranslator; import org.apache.commons.lang3.text.translate.EntityArrays; import org.apache.commons.lang3.text.translate.LookupTranslator; diff --git common/src/java/org/apache/hive/http/HttpServer.java common/src/java/org/apache/hive/http/HttpServer.java index 52253f94ac..51a2be2bdf 100644 --- common/src/java/org/apache/hive/http/HttpServer.java +++ common/src/java/org/apache/hive/http/HttpServer.java @@ -48,7 +48,7 @@ import com.google.common.base.Preconditions; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.math3.util.Pair; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.CommonConfigurationKeys; diff --git druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java index 430cc34501..1d7009b5af 100644 --- druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java +++ druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java @@ -718,7 +718,7 @@ static int getIntegerProperty(Table table, String propertyName, int defaultVal) } String[] vals = values.trim().split(","); for (String val : vals) { - if (org.apache.commons.lang.StringUtils.isNotBlank(val)) { + if (org.apache.commons.lang3.StringUtils.isNotBlank(val)) { rv.add(val); } } diff --git druid-handler/src/java/org/apache/hadoop/hive/druid/io/DruidOutputFormat.java druid-handler/src/java/org/apache/hadoop/hive/druid/io/DruidOutputFormat.java index 2ad6a7f8b0..6cf3ef2562 100644 --- druid-handler/src/java/org/apache/hadoop/hive/druid/io/DruidOutputFormat.java +++ druid-handler/src/java/org/apache/hadoop/hive/druid/io/DruidOutputFormat.java @@ -33,7 +33,7 @@ import org.apache.druid.segment.indexing.granularity.GranularitySpec; import org.apache.druid.segment.realtime.plumber.CustomVersioningPolicy; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.Constants; diff --git hbase-handler/src/java/org/apache/hadoop/hive/hbase/ColumnMappings.java hbase-handler/src/java/org/apache/hadoop/hive/hbase/ColumnMappings.java index f1887b5499..2f09662a4b 100644 --- hbase-handler/src/java/org/apache/hadoop/hive/hbase/ColumnMappings.java +++ hbase-handler/src/java/org/apache/hadoop/hive/hbase/ColumnMappings.java @@ -29,7 +29,7 @@ import java.util.List; import java.util.Properties; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.serde.serdeConstants; import org.apache.hadoop.hive.serde2.SerDeException; diff --git hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHFileOutputFormat.java hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHFileOutputFormat.java index 4fa0272951..415adca674 100644 --- hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHFileOutputFormat.java +++ hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHFileOutputFormat.java @@ -27,7 +27,7 @@ import java.util.SortedMap; import java.util.TreeMap; -import org.apache.commons.lang.NotImplementedException; +import org.apache.commons.lang3.NotImplementedException; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.conf.Configuration; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/HCatCli.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/HCatCli.java index 08e328a173..930e11ea8b 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/HCatCli.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/HCatCli.java @@ -36,7 +36,7 @@ import org.apache.commons.cli.Options; import org.apache.commons.cli.ParseException; import org.apache.commons.cli.Parser; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.io.SessionStream; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/SemanticAnalysis/CreateTableHook.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/SemanticAnalysis/CreateTableHook.java index 084bbfeee7..9b66e6be74 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/SemanticAnalysis/CreateTableHook.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/cli/SemanticAnalysis/CreateTableHook.java @@ -23,7 +23,7 @@ import java.util.List; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.ql.ddl.DDLDesc; @@ -72,7 +72,7 @@ public ASTNode preAnalyze(HiveSemanticAnalyzerHookContext context, for (int num = 1; num < numCh; num++) { ASTNode child = (ASTNode) ast.getChild(num); if (format.fillStorageFormat(child)) { - if (org.apache.commons.lang.StringUtils + if (org.apache.commons.lang3.StringUtils .isNotEmpty(format.getStorageHandler())) { return ast; } diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java index e851d2a6f4..b0b3276893 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java @@ -37,7 +37,7 @@ import com.google.common.collect.Maps; import org.apache.commons.codec.binary.Base64; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HiveClientCache.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HiveClientCache.java index e18dae983b..d786e3c482 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HiveClientCache.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HiveClientCache.java @@ -30,8 +30,8 @@ import javax.security.auth.login.LoginException; -import org.apache.commons.lang.builder.EqualsBuilder; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.EqualsBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.common.classification.InterfaceAudience; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.HiveMetaStoreClient; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatFieldSchema.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatFieldSchema.java index 350221528e..30af54f115 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatFieldSchema.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatFieldSchema.java @@ -20,7 +20,7 @@ import java.io.Serializable; -import org.apache.commons.lang.builder.ToStringBuilder; +import org.apache.commons.lang3.builder.ToStringBuilder; import org.apache.hadoop.hive.common.classification.InterfaceAudience; import org.apache.hadoop.hive.common.classification.InterfaceStability; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java index 4a76010904..5d0bef4a94 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java @@ -28,7 +28,7 @@ import java.util.Map; import java.util.Map.Entry; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatBaseInputFormat.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatBaseInputFormat.java index 002f63f7bb..076b4798ef 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatBaseInputFormat.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatBaseInputFormat.java @@ -338,7 +338,7 @@ private static InputJobInfo getJobInfo(Configuration conf) Iterator pathIterator = pathStrings.iterator(); while (pathIterator.hasNext()) { String pathString = pathIterator.next(); - if (ignoreInvalidPath && org.apache.commons.lang.StringUtils.isBlank(pathString)) { + if (ignoreInvalidPath && org.apache.commons.lang3.StringUtils.isBlank(pathString)) { continue; } Path path = new Path(pathString); diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/MultiOutputFormat.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/MultiOutputFormat.java index 8a080cf0a5..f1eed505a8 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/MultiOutputFormat.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/MultiOutputFormat.java @@ -32,7 +32,7 @@ import java.util.Map.Entry; import java.util.Set; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/TaskCommitContextRegistry.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/TaskCommitContextRegistry.java index 6dbb0a704c..9d83352df4 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/TaskCommitContextRegistry.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/TaskCommitContextRegistry.java @@ -19,7 +19,7 @@ package org.apache.hive.hcatalog.mapreduce; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.apache.hive.hcatalog.common.HCatConstants; import org.apache.hive.hcatalog.common.HCatUtil; diff --git hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java index 994c505d64..422342c947 100644 --- hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java +++ hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java @@ -31,7 +31,7 @@ import java.util.Map.Entry; import java.util.Properties; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.common.type.HiveChar; diff --git hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/messaging/jms/MessagingUtils.java hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/messaging/jms/MessagingUtils.java index 5ab5f513b2..8c8a8ce934 100644 --- hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/messaging/jms/MessagingUtils.java +++ hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/messaging/jms/MessagingUtils.java @@ -19,7 +19,7 @@ package org.apache.hive.hcatalog.messaging.jms; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hive.hcatalog.common.HCatConstants; import org.apache.hive.hcatalog.messaging.HCatEventMessage; import org.apache.hive.hcatalog.messaging.MessageFactory; diff --git hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/StrictRegexWriter.java hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/StrictRegexWriter.java index 6a9a47e85c..28406d38e8 100644 --- hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/StrictRegexWriter.java +++ hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/StrictRegexWriter.java @@ -23,7 +23,7 @@ import java.util.List; import java.util.Properties; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.Table; diff --git hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatClientHMSImpl.java hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatClientHMSImpl.java index 1cb35bbd06..b73b6fe524 100644 --- hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatClientHMSImpl.java +++ hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatClientHMSImpl.java @@ -28,7 +28,7 @@ import javax.annotation.Nullable; import org.apache.commons.lang3.tuple.Pair; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.classification.InterfaceAudience; import org.apache.hadoop.hive.common.classification.InterfaceStability; diff --git hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatTable.java hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatTable.java index ed2aef4758..ee60556a80 100644 --- hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatTable.java +++ hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatTable.java @@ -25,7 +25,7 @@ import java.util.Map; import com.google.common.collect.Maps; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.classification.InterfaceAudience; import org.apache.hadoop.hive.common.classification.InterfaceStability; diff --git hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/HcatDelegator.java hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/HcatDelegator.java index e6afd69563..ad447a7cf1 100644 --- hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/HcatDelegator.java +++ hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/HcatDelegator.java @@ -27,7 +27,7 @@ import javax.ws.rs.core.Response; import org.apache.commons.exec.ExecuteException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.FileStatus; diff --git hplsql/src/main/java/org/apache/hive/hplsql/Copy.java hplsql/src/main/java/org/apache/hive/hplsql/Copy.java index e562656282..d420e5ca17 100644 --- hplsql/src/main/java/org/apache/hive/hplsql/Copy.java +++ hplsql/src/main/java/org/apache/hive/hplsql/Copy.java @@ -35,7 +35,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hive.hplsql.Var; import org.antlr.v4.runtime.ParserRuleContext; -import org.apache.commons.lang.StringEscapeUtils; +import org.apache.commons.lang3.StringEscapeUtils; import org.apache.commons.lang3.tuple.Pair; public class Copy { diff --git hplsql/src/main/java/org/apache/hive/hplsql/functions/Function.java hplsql/src/main/java/org/apache/hive/hplsql/functions/Function.java index 279dd2048f..3554dac6ed 100644 --- hplsql/src/main/java/org/apache/hive/hplsql/functions/Function.java +++ hplsql/src/main/java/org/apache/hive/hplsql/functions/Function.java @@ -30,7 +30,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.antlr.v4.runtime.ParserRuleContext; import org.apache.hive.hplsql.*; diff --git hplsql/src/main/java/org/apache/hive/hplsql/functions/FunctionDatetime.java hplsql/src/main/java/org/apache/hive/hplsql/functions/FunctionDatetime.java index 373dd70da7..521c478af7 100644 --- hplsql/src/main/java/org/apache/hive/hplsql/functions/FunctionDatetime.java +++ hplsql/src/main/java/org/apache/hive/hplsql/functions/FunctionDatetime.java @@ -24,7 +24,7 @@ import java.util.Date; import java.util.TimeZone; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hive.hplsql.*; public class FunctionDatetime extends Function { diff --git itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java index 897a4014e9..43effeb64e 100644 --- itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java +++ itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.parse; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/session/TestClearDanglingScratchDir.java itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/session/TestClearDanglingScratchDir.java index c9bf4b59bd..82d3db5910 100644 --- itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/session/TestClearDanglingScratchDir.java +++ itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/session/TestClearDanglingScratchDir.java @@ -21,7 +21,7 @@ import java.io.PrintStream; import java.util.UUID; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.CommonConfigurationKeysPublic; import org.apache.hadoop.fs.Path; diff --git itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java index 18b5410dc4..04dd0ddad2 100644 --- itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java +++ itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java @@ -45,7 +45,7 @@ import com.google.common.base.Function; import com.google.common.collect.Lists; -import org.apache.commons.lang.exception.ExceptionUtils; +import org.apache.commons.lang3.exception.ExceptionUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; diff --git itests/hive-unit/src/test/java/org/apache/hive/beeline/schematool/TestSchemaTool.java itests/hive-unit/src/test/java/org/apache/hive/beeline/schematool/TestSchemaTool.java index e0b93f36ba..1f0006ce19 100644 --- itests/hive-unit/src/test/java/org/apache/hive/beeline/schematool/TestSchemaTool.java +++ itests/hive-unit/src/test/java/org/apache/hive/beeline/schematool/TestSchemaTool.java @@ -24,7 +24,7 @@ import java.io.IOException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.metastore.tools.schematool.HiveSchemaHelper; import org.apache.hadoop.hive.metastore.tools.schematool.HiveSchemaHelper.NestedScriptParser; import org.apache.hadoop.hive.metastore.tools.schematool.HiveSchemaHelper.PostgresCommandParser; diff --git itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniHS2.java itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniHS2.java index 03a1926440..a3299eefdb 100644 --- itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniHS2.java +++ itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniHS2.java @@ -58,7 +58,7 @@ import com.google.common.cache.Cache; import com.google.common.cache.CacheBuilder; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; diff --git itests/hive-unit/src/test/java/org/apache/hive/service/cli/operation/TestOperationLoggingLayout.java itests/hive-unit/src/test/java/org/apache/hive/service/cli/operation/TestOperationLoggingLayout.java index 6972ddef9b..d4d53d4d2d 100644 --- itests/hive-unit/src/test/java/org/apache/hive/service/cli/operation/TestOperationLoggingLayout.java +++ itests/hive-unit/src/test/java/org/apache/hive/service/cli/operation/TestOperationLoggingLayout.java @@ -24,6 +24,7 @@ import java.util.Map; import java.util.concurrent.ConcurrentHashMap; +import org.apache.hadoop.hive.common.LogUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.log.HushableRandomAccessFileAppender; import org.apache.hadoop.hive.ql.log.LogDivertAppender; @@ -132,22 +133,47 @@ private void appendHushableRandomAccessFileAppender(Appender queryAppender) { } } - @Test - public void testSwitchLogLayout() throws Exception { + private void executeWithOperationLog(String query, boolean queryLogEnabled) throws Exception { // verify whether the sql operation log is generated and fetch correctly. - OperationHandle operationHandle = client.executeStatement(sessionHandle, sqlCntStar, null); + OperationHandle operationHandle = client.executeStatement(sessionHandle, query, null); RowSet rowSetLog = client.fetchResults(operationHandle, FetchOrientation.FETCH_FIRST, 1000, - FetchType.LOG); - String queryId = getQueryId(rowSetLog); - Assert.assertNotNull("Could not find query id, perhaps a logging message changed", queryId); + FetchType.LOG); + String queryId = ""; + boolean expectedStopped = true; + if (queryLogEnabled) { + queryId = getQueryId(rowSetLog); + expectedStopped = false; + Assert.assertNotNull("Could not find query id, perhaps a logging message changed", queryId); + } else { + Assert.assertEquals("Operation log is generated even if query logging is disabled", rowSetLog.numRows(), 0); + Assert.assertNull("Query id present even if logging is disabled.", getQueryId(rowSetLog)); + } - checkAppenderState("before operation close ", LogDivertAppender.QUERY_ROUTING_APPENDER, queryId, false); - checkAppenderState("before operation close ", LogDivertAppenderForTest.TEST_QUERY_ROUTING_APPENDER, queryId, false); + checkAppenderState("before operation close ", LogDivertAppender.QUERY_ROUTING_APPENDER, queryId, expectedStopped); + checkAppenderState("before operation close ", LogDivertAppenderForTest.TEST_QUERY_ROUTING_APPENDER, queryId, expectedStopped); client.closeOperation(operationHandle); checkAppenderState("after operation close ", LogDivertAppender.QUERY_ROUTING_APPENDER, queryId, true); checkAppenderState("after operation close ", LogDivertAppenderForTest.TEST_QUERY_ROUTING_APPENDER, queryId, true); } + @Test + public void testSwitchLogLayout() throws Exception { + executeWithOperationLog(sqlCntStar, true); + } + + @Test + public void testQueryLogDisabled() throws Exception { + OperationHandle operationHandle = client.executeStatement(sessionHandle, + "set hive.server2.logging.operation.enabled=false", null); + client.closeOperation(operationHandle); + + executeWithOperationLog(sqlCntStar, false); + + operationHandle = client.executeStatement(sessionHandle, + "set hive.server2.logging.operation.enabled=true", null); + client.closeOperation(operationHandle); + } + @Test /** * Test to make sure that appending log event to HushableRandomAccessFileAppender even after diff --git itests/util/src/main/java/org/apache/hadoop/hive/cli/control/AbstractCliConfig.java itests/util/src/main/java/org/apache/hadoop/hive/cli/control/AbstractCliConfig.java index 0997681cbf..712af828fe 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/cli/control/AbstractCliConfig.java +++ itests/util/src/main/java/org/apache/hadoop/hive/cli/control/AbstractCliConfig.java @@ -32,7 +32,7 @@ import java.util.regex.Pattern; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.QTestSystemProperties; import org.apache.hadoop.hive.ql.QTestMiniClusters.FsType; import org.apache.hadoop.hive.ql.QTestMiniClusters.MiniClusterType; diff --git itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java index cd6bc34cc1..5b08f8b894 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java +++ itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java @@ -33,7 +33,7 @@ import java.util.stream.Stream; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConfUtil; import org.apache.hadoop.hive.ql.QTestProcessExecResult; diff --git itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseTestSetup.java itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseTestSetup.java index 75015808bc..8baf1464b9 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseTestSetup.java +++ itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseTestSetup.java @@ -21,7 +21,7 @@ import java.io.IOException; import java.util.Arrays; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseConfiguration; diff --git itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestResultProcessor.java itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestResultProcessor.java index 18b49fa7cb..204a9695b8 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestResultProcessor.java +++ itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestResultProcessor.java @@ -193,7 +193,7 @@ private static QTestProcessExecResult executeCmd(Collection args, String public static QTestProcessExecResult executeCmd(String[] args, String outFile, String errFile) throws Exception { - System.out.println("Running: " + org.apache.commons.lang.StringUtils.join(args, ' ')); + System.out.println("Running: " + org.apache.commons.lang3.StringUtils.join(args, ' ')); PrintStream out = outFile == null ? SessionState.getConsole().getChildOutStream() : new PrintStream(new FileOutputStream(outFile), true, "UTF-8"); diff --git itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestSyntaxUtil.java itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestSyntaxUtil.java index 3412cae427..c2f7acda08 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestSyntaxUtil.java +++ itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestSyntaxUtil.java @@ -22,7 +22,7 @@ import java.sql.SQLException; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.cli.CliSessionState; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.lockmgr.HiveTxnManager; diff --git itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java index 953253f16a..217049a2bf 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java +++ itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java @@ -41,7 +41,7 @@ import java.util.regex.Pattern; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckColumnAccessHook.java itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckColumnAccessHook.java index 138f4ae0d4..238db5b825 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckColumnAccessHook.java +++ itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckColumnAccessHook.java @@ -22,7 +22,7 @@ import java.util.Map; import java.util.LinkedHashMap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.QueryPlan; import org.apache.hadoop.hive.ql.session.SessionState; diff --git itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckTableAccessHook.java itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckTableAccessHook.java index 4e996df9a5..bcefe89b50 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckTableAccessHook.java +++ itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/CheckTableAccessHook.java @@ -21,7 +21,7 @@ import java.util.Map; import java.util.LinkedHashMap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.QueryPlan; import org.apache.hadoop.hive.ql.parse.TableAccessInfo; diff --git itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifySessionStateStackTracesHook.java itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifySessionStateStackTracesHook.java index aa75c481e8..6bb6a3c65b 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifySessionStateStackTracesHook.java +++ itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifySessionStateStackTracesHook.java @@ -20,7 +20,7 @@ import java.util.List; import java.util.Map.Entry; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.session.SessionState; import org.apache.hadoop.hive.ql.session.SessionState.LogHelper; diff --git itests/util/src/main/java/org/apache/hive/beeline/QFile.java itests/util/src/main/java/org/apache/hive/beeline/QFile.java index 34e7113683..aeb36f8f27 100644 --- itests/util/src/main/java/org/apache/hive/beeline/QFile.java +++ itests/util/src/main/java/org/apache/hive/beeline/QFile.java @@ -271,7 +271,7 @@ private QTestProcessExecResult executeDiff() throws IOException, InterruptedExce diffCommandArgs.add(getQuotedString(expectedOutputFile)); diffCommandArgs.add(getQuotedString(outputFile)); - System.out.println("Running: " + org.apache.commons.lang.StringUtils.join(diffCommandArgs, + System.out.println("Running: " + org.apache.commons.lang3.StringUtils.join(diffCommandArgs, ' ')); Process executor = Runtime.getRuntime().exec(diffCommandArgs.toArray( new String[diffCommandArgs.size()])); diff --git jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java index cb0b0d1c92..bc332c19f1 100644 --- jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java +++ jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java @@ -27,7 +27,7 @@ import org.apache.hive.service.rpc.thrift.TSetClientInfoResp; import org.apache.hive.service.rpc.thrift.TSetClientInfoReq; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.auth.HiveAuthUtils; import org.apache.hadoop.hive.shims.ShimLoader; import org.apache.hive.jdbc.Utils.JdbcConnectionParams; diff --git jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java index c6ac79373f..693203fab3 100644 --- jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java +++ jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java @@ -18,7 +18,7 @@ package org.apache.hive.jdbc; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.classification.InterfaceAudience.LimitedPrivate; import org.apache.hive.jdbc.logs.InPlaceUpdateStream; import org.apache.hive.service.cli.RowSet; diff --git llap-common/src/java/org/apache/hadoop/hive/llap/security/LlapTokenIdentifier.java llap-common/src/java/org/apache/hadoop/hive/llap/security/LlapTokenIdentifier.java index f2bb83a4b0..040932c870 100644 --- llap-common/src/java/org/apache/hadoop/hive/llap/security/LlapTokenIdentifier.java +++ llap-common/src/java/org/apache/hadoop/hive/llap/security/LlapTokenIdentifier.java @@ -22,7 +22,7 @@ import java.io.DataOutput; import java.io.IOException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.Text; import org.apache.hadoop.security.token.Token; diff --git llap-common/src/test/org/apache/hadoop/hive/llap/TestRow.java llap-common/src/test/org/apache/hadoop/hive/llap/TestRow.java index 37e934df3a..ed6b005345 100644 --- llap-common/src/test/org/apache/hadoop/hive/llap/TestRow.java +++ llap-common/src/test/org/apache/hadoop/hive/llap/TestRow.java @@ -21,7 +21,7 @@ import java.util.List; import java.util.Random; -import org.apache.commons.lang.RandomStringUtils; +import org.apache.commons.lang3.RandomStringUtils; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; diff --git llap-server/src/java/org/apache/hadoop/hive/llap/cache/LowLevelLrfuCachePolicy.java llap-server/src/java/org/apache/hadoop/hive/llap/cache/LowLevelLrfuCachePolicy.java index d1d6acd398..2afb899148 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/cache/LowLevelLrfuCachePolicy.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/cache/LowLevelLrfuCachePolicy.java @@ -21,7 +21,7 @@ import java.util.concurrent.atomic.AtomicLong; import java.util.concurrent.locks.ReentrantLock; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.llap.LlapUtil; diff --git llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapTaskReporter.java llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapTaskReporter.java index 33ade55ee1..89bf5713df 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapTaskReporter.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapTaskReporter.java @@ -33,7 +33,7 @@ import java.util.concurrent.locks.Condition; import java.util.concurrent.locks.ReentrantLock; -import org.apache.commons.lang.exception.ExceptionUtils; +import org.apache.commons.lang3.exception.ExceptionUtils; import org.apache.hadoop.hive.llap.counters.FragmentCountersMap; import org.apache.hadoop.hive.llap.counters.WmFragmentCounters; import org.apache.hadoop.hive.llap.daemon.SchedulerFragmentCompletingListener; diff --git llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java index 2ecb7a2830..d11bf1326c 100644 --- llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java +++ llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java @@ -67,7 +67,7 @@ import java.util.concurrent.locks.ReentrantReadWriteLock; import org.apache.commons.codec.binary.Base64; -import org.apache.commons.lang.mutable.MutableInt; +import org.apache.commons.lang3.mutable.MutableInt; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.JvmPauseMonitor; import org.apache.hadoop.hive.conf.HiveConf; diff --git metastore/src/java/org/apache/hadoop/hive/metastore/HiveClientCache.java metastore/src/java/org/apache/hadoop/hive/metastore/HiveClientCache.java index 6c33f63298..b68511d4fa 100644 --- metastore/src/java/org/apache/hadoop/hive/metastore/HiveClientCache.java +++ metastore/src/java/org/apache/hadoop/hive/metastore/HiveClientCache.java @@ -30,8 +30,8 @@ import javax.security.auth.login.LoginException; -import org.apache.commons.lang.builder.EqualsBuilder; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.EqualsBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.common.classification.InterfaceAudience; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.annotation.NoReconnect; diff --git metastore/src/java/org/apache/hadoop/hive/metastore/SerDeStorageSchemaReader.java metastore/src/java/org/apache/hadoop/hive/metastore/SerDeStorageSchemaReader.java index 59bcd5ca34..465d914043 100644 --- metastore/src/java/org/apache/hadoop/hive/metastore/SerDeStorageSchemaReader.java +++ metastore/src/java/org/apache/hadoop/hive/metastore/SerDeStorageSchemaReader.java @@ -35,11 +35,11 @@ try { if (envContext != null) { String addedJars = envContext.getProperties().get("hive.added.jars.path"); - if (org.apache.commons.lang.StringUtils.isNotBlank(addedJars)) { + if (org.apache.commons.lang3.StringUtils.isNotBlank(addedJars)) { //for thread safe orgHiveLoader = conf.getClassLoader(); ClassLoader loader = org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.addToClassPath( - orgHiveLoader, org.apache.commons.lang.StringUtils.split(addedJars, ",")); + orgHiveLoader, org.apache.commons.lang3.StringUtils.split(addedJars, ",")); conf.setClassLoader(loader); } } diff --git pom.xml pom.xml index c5e062b560..d4c23a0d8e 100644 --- pom.xml +++ pom.xml @@ -143,6 +143,7 @@ 1.4 2.6.1 2.6.0 + 1.8 10.14.1.0 3.1.0 0.1.2 @@ -373,6 +374,11 @@ junit ${junit.version} + + org.apache.commons + commons-text + ${commons-text.version} + org.apache.logging.log4j log4j-1.2-api @@ -1297,6 +1303,13 @@ true + + Do not use commons-lang + + org.apache.commons.lang.** + + true + diff --git ql/pom.xml ql/pom.xml index e5eed46947..08a8ff751b 100644 --- ql/pom.xml +++ ql/pom.xml @@ -119,6 +119,11 @@ commons-lang3 ${commons-lang3.version} + + org.apache.commons + commons-text + ${commons-text.version} + javolution javolution diff --git ql/src/java/org/apache/hadoop/hive/ql/Context.java ql/src/java/org/apache/hadoop/hive/ql/Context.java index 7a7fce1b0b..a85b94c475 100644 --- ql/src/java/org/apache/hadoop/hive/ql/Context.java +++ ql/src/java/org/apache/hadoop/hive/ql/Context.java @@ -923,7 +923,7 @@ public void resetStream() { * Little abbreviation for StringUtils. */ private static boolean strEquals(String str1, String str2) { - return org.apache.commons.lang.StringUtils.equals(str1, str2); + return org.apache.commons.lang3.StringUtils.equals(str1, str2); } /** diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/DDLUtils.java ql/src/java/org/apache/hadoop/hive/ql/ddl/DDLUtils.java index 3dc6bf56f2..eb8b858a24 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/DDLUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/DDLUtils.java @@ -28,7 +28,7 @@ import java.util.SortedMap; import java.util.TreeMap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/database/alter/location/AlterDatabaseSetLocationOperation.java ql/src/java/org/apache/hadoop/hive/ql/ddl/database/alter/location/AlterDatabaseSetLocationOperation.java index 44871b4c5d..748236cc1e 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/database/alter/location/AlterDatabaseSetLocationOperation.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/database/alter/location/AlterDatabaseSetLocationOperation.java @@ -22,7 +22,7 @@ import java.net.URISyntaxException; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.metastore.api.Database; import org.apache.hadoop.hive.ql.ErrorMsg; import org.apache.hadoop.hive.ql.ddl.DDLOperationContext; diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/function/desc/DescFunctionOperation.java ql/src/java/org/apache/hadoop/hive/ql/ddl/function/desc/DescFunctionOperation.java index 6a94a93ef9..ea3f522e58 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/function/desc/DescFunctionOperation.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/function/desc/DescFunctionOperation.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.exec.FunctionInfo.FunctionResource; -import static org.apache.commons.lang.StringUtils.join; +import static org.apache.commons.lang3.StringUtils.join; import java.io.DataOutputStream; import java.io.IOException; diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/table/AbstractAlterTableOperation.java ql/src/java/org/apache/hadoop/hive/ql/ddl/table/AbstractAlterTableOperation.java index 63bbaf9ecd..323cdc42d3 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/table/AbstractAlterTableOperation.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/table/AbstractAlterTableOperation.java @@ -22,7 +22,7 @@ import java.util.List; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.common.TableName; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/table/create/show/ShowCreateTableOperation.java ql/src/java/org/apache/hadoop/hive/ql/ddl/table/create/show/ShowCreateTableOperation.java index affed03fbb..9c584ae33f 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/table/create/show/ShowCreateTableOperation.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/table/create/show/ShowCreateTableOperation.java @@ -36,7 +36,7 @@ import java.util.Set; import java.util.SortedMap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.metastore.TableType; diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/DescTableOperation.java ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/DescTableOperation.java index 5178fb5fb5..7e467dd9f3 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/DescTableOperation.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/table/info/DescTableOperation.java @@ -26,7 +26,7 @@ import java.util.Map; import java.util.concurrent.TimeUnit; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.common.ValidTxnList; diff --git ql/src/java/org/apache/hadoop/hive/ql/ddl/table/misc/AlterTableSetPropertiesOperation.java ql/src/java/org/apache/hadoop/hive/ql/ddl/table/misc/AlterTableSetPropertiesOperation.java index 72db45755a..ff6b08b5d5 100644 --- ql/src/java/org/apache/hadoop/hive/ql/ddl/table/misc/AlterTableSetPropertiesOperation.java +++ ql/src/java/org/apache/hadoop/hive/ql/ddl/table/misc/AlterTableSetPropertiesOperation.java @@ -21,7 +21,7 @@ import java.util.ArrayList; import java.util.List; -import org.apache.commons.lang.BooleanUtils; +import org.apache.commons.lang3.BooleanUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/debug/Utils.java ql/src/java/org/apache/hadoop/hive/ql/debug/Utils.java index 4bdcea5df1..df952660a5 100644 --- ql/src/java/org/apache/hadoop/hive/ql/debug/Utils.java +++ ql/src/java/org/apache/hadoop/hive/ql/debug/Utils.java @@ -26,7 +26,7 @@ import javax.management.MBeanServer; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java index 3e2518bf69..bc42df121f 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java @@ -493,7 +493,7 @@ private String makeInputString(List dirs) { } private ValidWriteIdList extractValidWriteIdList() { - if (currDesc.getTableName() == null || !org.apache.commons.lang.StringUtils.isBlank(currDesc.getTableName())) { + if (currDesc.getTableName() == null || !org.apache.commons.lang3.StringUtils.isBlank(currDesc.getTableName())) { String txnString = job.get(ValidWriteIdList.VALID_WRITEIDS_KEY); LOG.debug("FetchOperator get writeIdStr: " + txnString); return txnString == null ? new ValidReaderWriteIdList() : new ValidReaderWriteIdList(txnString); diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java index 0643a54753..07b1fba8c6 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java @@ -26,7 +26,7 @@ import org.apache.commons.lang3.tuple.ImmutablePair; import org.apache.commons.lang3.tuple.Pair; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.Constants; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/OrcFileMergeOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/OrcFileMergeOperator.java index cdaad1b173..b53205a5a6 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/OrcFileMergeOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/OrcFileMergeOperator.java @@ -19,7 +19,7 @@ import java.io.IOException; -import org.apache.commons.lang.exception.ExceptionUtils; +import org.apache.commons.lang3.exception.ExceptionUtils; import org.apache.hadoop.hive.ql.CompilationOpContext; import org.apache.hadoop.hive.ql.io.orc.Writer; import org.apache.orc.TypeDescription; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java index aa62e90c0e..6c67bc7dd8 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java @@ -79,8 +79,7 @@ import org.apache.commons.collections.CollectionUtils; import org.apache.commons.collections.MapUtils; -import org.apache.commons.lang.StringUtils; -import org.apache.commons.lang.WordUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.StringEscapeUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.filecache.DistributedCache; @@ -260,6 +259,11 @@ @Deprecated protected static final String DEPRECATED_MAPRED_DFSCLIENT_PARALLELISM_MAX = "mapred.dfsclient.parallelism.max"; + // all common whitespaces as defined in Character.isWhitespace(char) + // Used primarily as a workaround until TEXT-175 is released + public static final char[] COMMON_WHITESPACE_CHARS = + { '\t', '\n', '\u000B', '\f', '\r', '\u001C', '\u001D', '\u001E', '\u001F', ' ' }; + private static final Object INPUT_SUMMARY_LOCK = new Object(); private static final Object ROOT_HDFS_DIR_LOCK = new Object(); @@ -865,7 +869,7 @@ public static String abbreviate(String str, int max) { String rev = StringUtils.reverse(str); // get the last few words - String suffix = WordUtils.abbreviate(rev, 0, suffixlength, StringUtils.EMPTY); + String suffix = StringUtils.abbreviate(rev, suffixlength); suffix = StringUtils.reverse(suffix); // first few .. diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecDriver.java ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecDriver.java index 196ca39689..46764d4e30 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecDriver.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecDriver.java @@ -33,7 +33,7 @@ import java.util.Properties; import java.util.concurrent.ThreadLocalRandom; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.AddToClassPathAction; import org.apache.hadoop.hive.ql.exec.SerializationUtilities; import org.apache.hadoop.hive.ql.log.LogDivertAppenderForTest; @@ -438,7 +438,7 @@ public int execute() { } // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils console.printError(mesg, "\n" + org.apache.hadoop.util.StringUtils.stringifyException(e)); success = false; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java index dffc5cbd45..b7bb7d0af6 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java @@ -27,7 +27,7 @@ import java.util.Properties; import org.apache.commons.io.IOUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.ContentSummary; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileUtil; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java index 9d499c093f..78e3fd57f4 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java @@ -37,7 +37,7 @@ import java.util.Properties; import org.apache.commons.io.IOUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/BytesBytesMultiHashMap.java ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/BytesBytesMultiHashMap.java index dbf75b426d..999fe3d9b5 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/BytesBytesMultiHashMap.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/BytesBytesMultiHashMap.java @@ -23,7 +23,7 @@ import java.util.Map; import java.util.TreeMap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.MemoryEstimate; import org.apache.hadoop.hive.ql.util.JavaDataModel; import org.slf4j.Logger; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadTask.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadTask.java index 7e5a40ad17..a2c467bafd 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadTask.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadTask.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.exec.repl; import com.google.common.collect.Collections2; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadWork.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadWork.java index 44d2535a44..272373e273 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadWork.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplLoadWork.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.exec.repl; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.repl.ReplScope; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/BootstrapEventsIterator.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/BootstrapEventsIterator.java index ab6e09f2df..1af6a4c637 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/BootstrapEventsIterator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/BootstrapEventsIterator.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.exec.repl.bootstrap.events.filesystem; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.*; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.repl.bootstrap.load.ReplicationState; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/DatabaseEventsIterator.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/DatabaseEventsIterator.java index 5665bda537..4c84797efe 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/DatabaseEventsIterator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/DatabaseEventsIterator.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.exec.repl.bootstrap.events.filesystem; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.LocatedFileStatus; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSDatabaseEvent.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSDatabaseEvent.java index 0e91e25882..16e4128a5d 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSDatabaseEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSDatabaseEvent.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.exec.repl.bootstrap.events.filesystem; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSTableEvent.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSTableEvent.java index b90f609a29..6d38c0309f 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSTableEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/events/filesystem/FSTableEvent.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.exec.repl.bootstrap.events.filesystem; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/load/LoadFunction.java ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/load/LoadFunction.java index 19e1a8b0eb..8815eeebe1 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/load/LoadFunction.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/repl/bootstrap/load/LoadFunction.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.exec.repl.bootstrap.load; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.metastore.api.Function; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java index f23db78929..95feb7c64a 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java @@ -32,7 +32,7 @@ import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.io.NullScanFileSystem; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlan.java ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlan.java index 8244dcb1a9..2f6dd4a4de 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlan.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlan.java @@ -28,7 +28,7 @@ import java.util.Map; import java.util.Set; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.ExplainTask; import org.apache.hadoop.hive.ql.exec.Utilities; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/LocalSparkJobMonitor.java ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/LocalSparkJobMonitor.java index aeef3c11ce..911699d92c 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/LocalSparkJobMonitor.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/LocalSparkJobMonitor.java @@ -123,7 +123,7 @@ public int startMonitor() { msg = "Failed to monitor Job[ " + sparkJobStatus.getJobId() + "]" + msg; // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils LOG.error(msg, e); console.printError(msg, "\n" + org.apache.hadoop.util.StringUtils.stringifyException(e)); rc = 1; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java index 87b69cbae4..35d20e3439 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java @@ -174,7 +174,7 @@ public int startMonitor() { msg = "Failed to monitor Job[" + sparkJobStatus.getJobId() + "]" + msg; // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils console.printError(msg, "\n" + org.apache.hadoop.util.StringUtils.stringifyException(e)); } rc = 1; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java index 94ef91232f..067c0f002b 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java @@ -50,7 +50,7 @@ import java.util.zip.ZipOutputStream; import org.apache.commons.io.FilenameUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.tez.mapreduce.common.MRInputSplitDistributor; import org.apache.tez.mapreduce.hadoop.InputSplitInfo; import org.apache.tez.mapreduce.protos.MRRuntimeProtos; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/tez/Utils.java ql/src/java/org/apache/hadoop/hive/ql/exec/tez/Utils.java index db1a0e6387..cc9c4ce8af 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/tez/Utils.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/tez/Utils.java @@ -23,7 +23,7 @@ import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.llap.registry.LlapServiceInstance; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/tez/WorkloadManager.java ql/src/java/org/apache/hadoop/hive/ql/exec/tez/WorkloadManager.java index 37c3714017..1aa133e517 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/tez/WorkloadManager.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/tez/WorkloadManager.java @@ -54,7 +54,7 @@ import java.util.concurrent.locks.Condition; import java.util.concurrent.locks.ReentrantLock; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnOrderedMap.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnOrderedMap.java index 0ec3838af6..0511b99d13 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnOrderedMap.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnOrderedMap.java @@ -22,7 +22,7 @@ import java.util.Map; import java.util.TreeMap; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorGroupByOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorGroupByOperator.java index 71d015b631..7fe6151a17 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorGroupByOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorGroupByOperator.java @@ -29,7 +29,7 @@ import java.util.List; import java.util.Map; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.type.DataTypePhysicalVariation; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java index ae2edf4fd7..1441fe5dcd 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java @@ -34,7 +34,7 @@ import java.util.TreeSet; import java.util.regex.Pattern; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.ql.exec.vector.expressions.BucketNumExpression; import org.apache.hadoop.hive.ql.exec.vector.expressions.CastBooleanToCharViaLongToChar; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java index ebd89e6bb9..6cccd9e8b2 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java @@ -23,8 +23,8 @@ import java.util.Arrays; import java.util.List; -import org.apache.commons.lang.ArrayUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.ArrayUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.DataTypePhysicalVariation; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.common.type.HiveChar; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedRowBatchCtx.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedRowBatchCtx.java index 83e41a37cc..ebffbc1a06 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedRowBatchCtx.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedRowBatchCtx.java @@ -348,7 +348,7 @@ public VectorizedRowBatch createVectorizedRowBatch() final int partitionEndColumnNum = dataColumnCount + partitionColumnCount; for (int partitionColumnNum = dataColumnCount; partitionColumnNum < partitionEndColumnNum; partitionColumnNum++) { result.cols[partitionColumnNum] = - VectorizedBatchUtil.createColumnVector(rowColumnTypeInfos[partitionColumnNum]); + createColumnVectorFromRowColumnTypeInfos(partitionColumnNum); } final int virtualEndColumnNum = partitionEndColumnNum + virtualColumnCount; for (int virtualColumnNum = partitionEndColumnNum; virtualColumnNum < virtualEndColumnNum; virtualColumnNum++) { @@ -527,13 +527,27 @@ public void addPartitionColsToBatch(ColumnVector[] cols, Object[] partitionValue break; case DECIMAL: { - DecimalColumnVector dv = (DecimalColumnVector) cols[colIndex]; - if (value == null) { - dv.noNulls = false; - dv.isNull[0] = true; - dv.isRepeating = true; + DataTypePhysicalVariation dataTypePhysicalVariation = rowDataTypePhysicalVariations != null ? + rowDataTypePhysicalVariations[colIndex] : DataTypePhysicalVariation.NONE; + + if (dataTypePhysicalVariation == DataTypePhysicalVariation.DECIMAL_64) { + Decimal64ColumnVector dv = (Decimal64ColumnVector) cols[colIndex]; + if (value == null) { + dv.noNulls = false; + dv.isNull[0] = true; + dv.isRepeating = true; + } else { + dv.fill(((HiveDecimal) value).longValue()); + } } else { - dv.fill((HiveDecimal) value); + DecimalColumnVector dv = (DecimalColumnVector) cols[colIndex]; + if (value == null) { + dv.noNulls = false; + dv.isNull[0] = true; + dv.isRepeating = true; + } else { + dv.fill((HiveDecimal) value); + } } } break; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/AbstractFilterStringColLikeStringScalar.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/AbstractFilterStringColLikeStringScalar.java index aa7228689b..85c07b6dc5 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/AbstractFilterStringColLikeStringScalar.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/AbstractFilterStringColLikeStringScalar.java @@ -29,7 +29,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringInitCap.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringInitCap.java index b30da5631d..346c5a1b5c 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringInitCap.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringInitCap.java @@ -18,7 +18,8 @@ package org.apache.hadoop.hive.ql.exec.vector.expressions; -import org.apache.commons.lang.WordUtils; +import org.apache.commons.text.WordUtils; +import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.io.Text; /** @@ -40,7 +41,7 @@ public Text evaluate(Text s) { if (s == null) { return null; } - t.set(WordUtils.capitalizeFully(s.toString())); + t.set(WordUtils.capitalizeFully(s.toString(), Utilities.COMMON_WHITESPACE_CHARS)); return t; } }); @@ -49,4 +50,4 @@ public Text evaluate(Text s) { public StringInitCap() { super(); } -} \ No newline at end of file +} diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStart.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStart.java index 619c603198..3236439ecf 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStart.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStart.java @@ -21,7 +21,7 @@ import java.nio.charset.StandardCharsets; import java.util.Arrays; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor; import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStartLen.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStartLen.java index f6f10c5ece..c23d451298 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStartLen.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/StringSubstrColStartLen.java @@ -21,7 +21,7 @@ import java.nio.charset.StandardCharsets; import java.util.Arrays; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor; import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/VectorExpressionWriterFactory.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/VectorExpressionWriterFactory.java index 7829b221e5..163347febe 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/VectorExpressionWriterFactory.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/VectorExpressionWriterFactory.java @@ -25,8 +25,8 @@ import java.util.List; import java.util.Map; -import org.apache.commons.lang.ArrayUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.ArrayUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.HiveChar; import org.apache.hadoop.hive.common.type.HiveDecimal; import org.apache.hadoop.hive.common.type.HiveIntervalDayTime; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinCommonOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinCommonOperator.java index 551d362185..fdbfe225b2 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinCommonOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinCommonOperator.java @@ -23,7 +23,7 @@ import java.util.List; import java.util.Map.Entry; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.hive.common.type.DataTypePhysicalVariation; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinGenerateResultOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinGenerateResultOperator.java index f5bb547493..66db5e65ca 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinGenerateResultOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/VectorMapJoinGenerateResultOperator.java @@ -22,7 +22,7 @@ import java.util.ArrayList; import java.util.List; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.conf.Configuration; diff --git ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageInfo.java ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageInfo.java index 532c7d0b60..f0a704d26a 100644 --- ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageInfo.java +++ ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageInfo.java @@ -26,7 +26,7 @@ import java.util.Set; import org.apache.commons.collections.SetUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.StringInternUtils; import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.FieldSchema; diff --git ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java index ec4c04c54b..0160fcabb9 100644 --- ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java +++ ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java @@ -25,7 +25,7 @@ import com.google.gson.stream.JsonWriter; import org.apache.commons.collections.SetUtils; import org.apache.commons.io.output.StringBuilderWriter; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.tuple.Pair; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.Warehouse; diff --git ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java index 14d18844b3..ed68e0f4af 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java @@ -1633,6 +1633,9 @@ private static boolean isValidBase(ParsedBase parsedBase, ValidWriteIdList write //By definition there are no open txns with id < 1. return true; } + if (writeIdList.getMinOpenWriteId() != null && parsedBase.getWriteId() <= writeIdList.getMinOpenWriteId()) { + return true; + } if(isCompactedBase(parsedBase, fs, (HdfsDirSnapshot) null)) { return writeIdList.isValidBase(parsedBase.getWriteId()); } @@ -2843,8 +2846,13 @@ private static boolean isLockableTable(Table t) { HiveConf conf) { List lockComponents = new ArrayList<>(); // For each source to read, get a shared lock + boolean skipReadLock = !conf.getBoolVar(ConfVars.HIVE_TXN_READ_LOCKS); for (ReadEntity input : inputs) { - if (input.isDummy() || !input.needsLock() || input.isUpdateOrDelete() || !AcidUtils.needsLock(input)) { + if (input.isDummy() + || !input.needsLock() + || input.isUpdateOrDelete() + || !AcidUtils.needsLock(input) + || skipReadLock) { // We don't want to acquire read locks during update or delete as we'll be acquiring write // locks instead. Also, there's no need to lock temp tables since they're session wide continue; diff --git ql/src/java/org/apache/hadoop/hive/ql/io/RCFileOutputFormat.java ql/src/java/org/apache/hadoop/hive/ql/io/RCFileOutputFormat.java index 98ed43864f..37aeba0a82 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/RCFileOutputFormat.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/RCFileOutputFormat.java @@ -21,7 +21,7 @@ import java.io.IOException; import java.util.Properties; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java index 39f77edcd9..7fb3878ee6 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java @@ -20,7 +20,7 @@ import org.apache.hadoop.hive.ql.exec.mr.ExecDriver; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; @@ -167,7 +167,7 @@ public int execute() { } // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils console.printError(mesg, "\n" + org.apache.hadoop.util.StringUtils.stringifyException(e)); diff --git ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java index ae9f0225e9..0458c946c0 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java @@ -22,7 +22,7 @@ import java.io.Serializable; import java.util.concurrent.ThreadLocalRandom; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.JavaUtils; @@ -192,7 +192,7 @@ public int execute() { } catch (Exception e) { String mesg = rj != null ? ("Ended Job = " + rj.getJobID()) : "Job Submission failed"; // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils LOG.error(mesg, org.apache.hadoop.util.StringUtils.stringifyException(e)); setException(e); diff --git ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java index 286a47f150..08aeeb2acd 100644 --- ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java +++ ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java @@ -20,8 +20,8 @@ import java.util.Map; -import org.apache.commons.lang.StringUtils; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.StringUtils; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.common.StringInternUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.metadata.DummyPartition; diff --git ql/src/java/org/apache/hadoop/hive/ql/lockmgr/zookeeper/ZooKeeperHiveLock.java ql/src/java/org/apache/hadoop/hive/ql/lockmgr/zookeeper/ZooKeeperHiveLock.java index d295c15f3f..fe81e0400a 100644 --- ql/src/java/org/apache/hadoop/hive/ql/lockmgr/zookeeper/ZooKeeperHiveLock.java +++ ql/src/java/org/apache/hadoop/hive/ql/lockmgr/zookeeper/ZooKeeperHiveLock.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.lockmgr.zookeeper; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.ql.lockmgr.HiveLock; import org.apache.hadoop.hive.ql.lockmgr.HiveLockMode; import org.apache.hadoop.hive.ql.lockmgr.HiveLockObject; diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java index 75a0ea5d19..7f061d4a6b 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java @@ -1923,7 +1923,7 @@ public static Boolean isOutdatedMaterializedView(Table materializedViewTable, fi long defaultTimeWindow, List tablesUsed, boolean forceMVContentsUpToDate) { // Check if materialization defined its own invalidation time window String timeWindowString = materializedViewTable.getProperty(MATERIALIZED_VIEW_REWRITING_TIME_WINDOW); - long timeWindow = org.apache.commons.lang.StringUtils.isEmpty(timeWindowString) ? defaultTimeWindow : + long timeWindow = org.apache.commons.lang3.StringUtils.isEmpty(timeWindowString) ? defaultTimeWindow : HiveConf.toTime(timeWindowString, HiveConf.getDefaultTimeUnit(HiveConf.ConfVars.HIVE_MATERIALIZED_VIEW_REWRITING_TIME_WINDOW), TimeUnit.MILLISECONDS); diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/JarUtils.java ql/src/java/org/apache/hadoop/hive/ql/metadata/JarUtils.java index 113a6e36eb..cde7459170 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/JarUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/JarUtils.java @@ -41,7 +41,7 @@ import java.util.zip.ZipOutputStream; import org.apache.commons.io.IOUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/SessionHiveMetaStoreClient.java ql/src/java/org/apache/hadoop/hive/ql/metadata/SessionHiveMetaStoreClient.java index af77c4e9b9..3dcf876af3 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/SessionHiveMetaStoreClient.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/SessionHiveMetaStoreClient.java @@ -1511,7 +1511,7 @@ private PartitionSpecProxy getPartitionSpecProxy(org.apache.hadoop.hive.metastor private String generateJDOFilter(org.apache.hadoop.hive.metastore.api.Table table, String filter) throws MetaException { - ExpressionTree exprTree = org.apache.commons.lang.StringUtils.isNotEmpty(filter) + ExpressionTree exprTree = org.apache.commons.lang3.StringUtils.isNotEmpty(filter) ? PartFilterExprUtil.getFilterParser(filter).tree : ExpressionTree.EMPTY_TREE; return generateJDOFilter(table, exprTree); } diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/JsonMetaDataFormatter.java ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/JsonMetaDataFormatter.java index bff163fdec..8d57604190 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/JsonMetaDataFormatter.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/JsonMetaDataFormatter.java @@ -32,7 +32,7 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.io.IOUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -172,7 +172,7 @@ public void showMaterializedViews(DataOutputStream out, List materialized final String refreshMode = "Manual refresh"; final String timeWindowString = mv.getProperty(MATERIALIZED_VIEW_REWRITING_TIME_WINDOW); final String mode; - if (!org.apache.commons.lang.StringUtils.isEmpty(timeWindowString)) { + if (!org.apache.commons.lang3.StringUtils.isEmpty(timeWindowString)) { long time = HiveConf.toTime(timeWindowString, HiveConf.getDefaultTimeUnit(HiveConf.ConfVars.HIVE_MATERIALIZED_VIEW_REWRITING_TIME_WINDOW), TimeUnit.MINUTES); diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java index fef02934ef..5dbb69d302 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.metadata.formatting; import org.apache.commons.collections.CollectionUtils; -import org.apache.commons.lang.StringEscapeUtils; +import org.apache.commons.lang3.StringEscapeUtils; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.common.type.HiveDecimal; import org.apache.hadoop.hive.common.type.Timestamp; diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java index bb17a14c52..6b35581357 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java @@ -33,7 +33,7 @@ import org.apache.hive.common.util.HiveStringUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import org.apache.commons.lang.StringEscapeUtils; +import org.apache.commons.lang3.StringEscapeUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -179,7 +179,7 @@ public void showMaterializedViews(DataOutputStream out, List
materialized final String refreshMode = "Manual refresh"; final String timeWindowString = mv.getProperty(MATERIALIZED_VIEW_REWRITING_TIME_WINDOW); final String mode; - if (!org.apache.commons.lang.StringUtils.isEmpty(timeWindowString)) { + if (!org.apache.commons.lang3.StringUtils.isEmpty(timeWindowString)) { long time = HiveConf.toTime(timeWindowString, HiveConf.getDefaultTimeUnit(HiveConf.ConfVars.HIVE_MATERIALIZED_VIEW_REWRITING_TIME_WINDOW), TimeUnit.MINUTES); diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConstantPropagateProcFactory.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConstantPropagateProcFactory.java index 69ee9af629..f6a8e910de 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConstantPropagateProcFactory.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConstantPropagateProcFactory.java @@ -28,7 +28,7 @@ import java.util.Set; import java.util.Stack; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.ColumnInfo; import org.apache.hadoop.hive.ql.exec.FileSinkOperator; diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/SharedWorkOptimizer.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/SharedWorkOptimizer.java index f6b43dbb3d..2f49985800 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/SharedWorkOptimizer.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/SharedWorkOptimizer.java @@ -36,7 +36,7 @@ import java.util.TreeMap; import java.util.stream.Collectors; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.ql.exec.AppMasterEventOperator; import org.apache.hadoop.hive.ql.exec.DummyStoreOperator; diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/SimpleFetchOptimizer.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/SimpleFetchOptimizer.java index 89f6ee1617..fe25ca4071 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/SimpleFetchOptimizer.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/SimpleFetchOptimizer.java @@ -117,7 +117,7 @@ public ParseContext transform(ParseContext pctx) throws SemanticException { } } catch (Exception e) { // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils LOG.error(org.apache.hadoop.util.StringUtils.stringifyException(e)); if (e instanceof SemanticException) { throw (SemanticException) e; diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedDynPartitionTimeGranularityOptimizer.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedDynPartitionTimeGranularityOptimizer.java index 4297537adb..384c857fa1 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedDynPartitionTimeGranularityOptimizer.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedDynPartitionTimeGranularityOptimizer.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.optimizer; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.Constants; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.FieldSchema; diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePartitionPruneRule.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePartitionPruneRule.java index 39973d884b..28f097aa49 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePartitionPruneRule.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePartitionPruneRule.java @@ -22,7 +22,7 @@ import org.apache.calcite.rel.core.Filter; import org.apache.calcite.rex.RexNode; import org.apache.calcite.util.Pair; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.optimizer.calcite.RelOptHiveTable; import org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveFilter; diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/LineageCtx.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/LineageCtx.java index 58b90d8cc3..83c30446e0 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/LineageCtx.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/LineageCtx.java @@ -25,7 +25,7 @@ import java.util.Map; import java.util.Set; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.tuple.Pair; import org.apache.hadoop.hive.ql.exec.ColumnInfo; import org.apache.hadoop.hive.ql.exec.FileSinkOperator; diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java index ab2e1e1d20..ef20520016 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java @@ -62,7 +62,7 @@ protected void generatePredicate(NodeProcessorCtx procCtx, FilterOperator fop, prunedPartList = PartitionPruner.prune(top, parseCtx, alias); } catch (HiveException e) { // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils throw new SemanticException(e.getMessage(), e); } diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java index 82da3eaf9a..20d478324c 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java @@ -110,7 +110,7 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, prunedPartList = pctx.getPrunedPartitions(alias, top); } catch (HiveException e) { // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils throw new SemanticException(e.getMessage(), e); } diff --git ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java index 73e22f3145..8fc834efb4 100644 --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java @@ -40,7 +40,7 @@ import java.util.TreeSet; import java.util.regex.Pattern; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.commons.lang3.tuple.ImmutablePair; import org.apache.hadoop.hive.ql.exec.vector.VectorizedInputFormatInterface; import org.apache.hadoop.hive.ql.exec.vector.expressions.ConvertDecimal64ToDecimal; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java index 8aae641e61..f588b0d065 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/EximUtil.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.parse; import org.apache.commons.io.IOUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java index e9324ee4f3..7b6ce10520 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java @@ -19,8 +19,8 @@ package org.apache.hadoop.hive.ql.parse; import org.antlr.runtime.tree.Tree; -import org.apache.commons.lang.ObjectUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.ObjectUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java index d1fdfc098f..94923e3f89 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java @@ -32,7 +32,7 @@ import org.antlr.runtime.tree.Tree; import org.apache.commons.codec.DecoderException; import org.apache.commons.codec.net.URLCodec; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -212,7 +212,7 @@ private URI initializeFromURI(String fromPath, boolean isLocal) } } catch (IOException e) { // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils throw new SemanticException(ErrorMsg.INVALID_PATH.getMsg(fromTree), e); } diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/MergeSemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/MergeSemanticAnalyzer.java index db6d5519e3..13f9243672 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/MergeSemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/MergeSemanticAnalyzer.java @@ -27,7 +27,7 @@ import org.antlr.runtime.TokenRewriteStream; import org.apache.commons.collections.MapUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.Warehouse; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java index ffaf4ee966..d68ce1d8a7 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.parse; import org.antlr.runtime.tree.Tree; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.ValidTxnList; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index f5df9cd302..02d25efa1e 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -60,7 +60,7 @@ import org.apache.calcite.rel.RelNode; import org.apache.calcite.util.ImmutableBitSet; import org.apache.commons.collections.CollectionUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.tuple.Pair; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileStatus; @@ -2032,7 +2032,7 @@ private void getMaterializationMetadata(QB qb) throws SemanticException { } } catch (HiveException e) { // Has to use full name to make sure it does not conflict with - // org.apache.commons.lang.StringUtils + // org.apache.commons.lang3.StringUtils LOG.error(org.apache.hadoop.util.StringUtils.stringifyException(e)); if (e instanceof SemanticException) { throw (SemanticException)e; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/StorageFormat.java ql/src/java/org/apache/hadoop/hive/ql/parse/StorageFormat.java index edf582a6af..2350646c36 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/StorageFormat.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/StorageFormat.java @@ -21,7 +21,7 @@ import java.util.HashMap; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.io.IOConstants; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java index e0f4826ed7..c4fe974411 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java @@ -29,7 +29,7 @@ import java.util.Stack; import org.apache.calcite.rel.RelNode; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.math.NumberUtils; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.common.type.HiveChar; diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CreateDatabaseHandler.java ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CreateDatabaseHandler.java index cc30131bdf..42fa88c5fb 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CreateDatabaseHandler.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CreateDatabaseHandler.java @@ -18,7 +18,7 @@ */ package org.apache.hadoop.hive.ql.parse.repl.load.message; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.metastore.api.Database; diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/ExplainLockDesc.java ql/src/java/org/apache/hadoop/hive/ql/plan/ExplainLockDesc.java index 1e7fb2eab8..af127053b6 100644 --- ql/src/java/org/apache/hadoop/hive/ql/plan/ExplainLockDesc.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/ExplainLockDesc.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.plan; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.metastore.api.LockComponent; import org.apache.hadoop.hive.metastore.api.LockType; diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeColumnDesc.java ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeColumnDesc.java index 671fc28eda..b82fef6d79 100755 --- ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeColumnDesc.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeColumnDesc.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.plan; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.ql.exec.ColumnInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java index e3324ca220..f32fd4a4a9 100755 --- ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java @@ -21,7 +21,7 @@ import java.io.Serializable; import java.util.List; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.common.StringInternUtils; import org.apache.hadoop.hive.serde.serdeConstants; import org.apache.hadoop.hive.serde2.objectinspector.ConstantObjectInspector; diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeFieldDesc.java ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeFieldDesc.java index 8e0cd315ab..415dc5d234 100755 --- ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeFieldDesc.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeFieldDesc.java @@ -24,7 +24,7 @@ import java.util.LinkedHashSet; import java.util.List; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; /** diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java index 002aef6ad2..12b0a5be16 100644 --- ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java @@ -25,7 +25,7 @@ import org.apache.commons.collections.Bag; import org.apache.commons.collections.bag.TreeBag; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/VectorPTFDesc.java ql/src/java/org/apache/hadoop/hive/ql/plan/VectorPTFDesc.java index 54efca8b79..4fba07508f 100644 --- ql/src/java/org/apache/hadoop/hive/ql/plan/VectorPTFDesc.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/VectorPTFDesc.java @@ -23,7 +23,7 @@ import java.util.List; import java.util.TreeSet; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.hive.ql.exec.vector.ColumnVector.Type; import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression; import org.apache.hadoop.hive.ql.exec.vector.ptf.VectorPTFEvaluatorBase; diff --git ql/src/java/org/apache/hadoop/hive/ql/plan/ptf/PTFInputDef.java ql/src/java/org/apache/hadoop/hive/ql/plan/ptf/PTFInputDef.java index bbe683b134..76559ab265 100644 --- ql/src/java/org/apache/hadoop/hive/ql/plan/ptf/PTFInputDef.java +++ ql/src/java/org/apache/hadoop/hive/ql/plan/ptf/PTFInputDef.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.plan.ptf; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.RowSchema; import org.apache.hadoop.hive.ql.plan.Explain; diff --git ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java index ee57e73b3b..7ba813df4f 100644 --- ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java +++ ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java @@ -21,7 +21,7 @@ import java.util.Arrays; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveVariableSource; import org.apache.hadoop.hive.conf.VariableSubstitution; import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveOperationType; diff --git ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessorFactory.java ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessorFactory.java index 5302800f88..977ab5372d 100644 --- ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessorFactory.java +++ ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessorFactory.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.processors; -import static org.apache.commons.lang.StringUtils.isBlank; +import static org.apache.commons.lang3.StringUtils.isBlank; import java.io.IOException; import java.sql.SQLException; diff --git ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java index c42fb050d8..c0a35d2720 100644 --- ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java +++ ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java @@ -21,7 +21,7 @@ import java.util.Arrays; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveVariableSource; import org.apache.hadoop.hive.conf.VariableSubstitution; import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveOperationType; diff --git ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java index 69439f14cb..ced9876c22 100644 --- ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java +++ ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.processors; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.Schema; import org.apache.hadoop.hive.ql.session.SessionState; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AddPartitionEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AddPartitionEvent.java index 1cb33c011b..54409dbd66 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AddPartitionEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AddPartitionEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.FieldSchema; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterDatabaseEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterDatabaseEvent.java index e78f27b479..0889fc61c0 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterDatabaseEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterDatabaseEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Database; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterPartitionEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterPartitionEvent.java index f83a73770f..f16a36b81d 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterPartitionEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterPartitionEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Partition; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterTableEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterTableEvent.java index bb688d6b96..5917b04561 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterTableEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/AlterTableEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Table; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateDatabaseEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateDatabaseEvent.java index 969740ce88..017b6c2ba5 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateDatabaseEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateDatabaseEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Database; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateTableEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateTableEvent.java index f3fa7586a5..c7f73bbfb4 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateTableEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/CreateTableEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Table; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropDatabaseEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropDatabaseEvent.java index c642378ffa..f078cdb6ca 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropDatabaseEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropDatabaseEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Database; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropPartitionEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropPartitionEvent.java index c24988774f..8f10ec5ac5 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropPartitionEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropPartitionEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.FieldSchema; diff --git ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropTableEvent.java ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropTableEvent.java index 393a0ce93a..78fdb00bec 100644 --- ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropTableEvent.java +++ ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/metastore/events/DropTableEvent.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin.metastore.events; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.metastore.api.Table; diff --git ql/src/java/org/apache/hadoop/hive/ql/session/OperationLog.java ql/src/java/org/apache/hadoop/hive/ql/session/OperationLog.java index c48dc42091..460f54ed93 100644 --- ql/src/java/org/apache/hadoop/hive/ql/session/OperationLog.java +++ ql/src/java/org/apache/hadoop/hive/ql/session/OperationLog.java @@ -31,7 +31,7 @@ import org.apache.commons.collections.CollectionUtils; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.io.IOUtils; import org.slf4j.Logger; diff --git ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java index de6aebbe59..072758bdcd 100644 --- ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java +++ ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java @@ -47,7 +47,7 @@ import java.util.concurrent.atomic.AtomicLong; import java.util.concurrent.locks.ReentrantLock; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; diff --git ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/MinorQueryCompactor.java ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/MinorQueryCompactor.java index 0eee0161a8..f96a0481b8 100644 --- ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/MinorQueryCompactor.java +++ ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/MinorQueryCompactor.java @@ -17,8 +17,8 @@ */ package org.apache.hadoop.hive.ql.txn.compactor; -import org.apache.commons.lang.ArrayUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.ArrayUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.ValidWriteIdList; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/QueryCompactor.java ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/QueryCompactor.java index 9896df3bb6..3097d1bbdd 100644 --- ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/QueryCompactor.java +++ ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/QueryCompactor.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.txn.compactor; -import org.apache.commons.lang.exception.ExceptionUtils; +import org.apache.commons.lang3.exception.ExceptionUtils; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseTrim.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseTrim.java index 95d91ddc58..7d1e9c025c 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseTrim.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseTrim.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.udf.generic; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.UDFArgumentException; import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java index c5f2a51a95..248f6da026 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.udf.generic; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.UDFArgumentException; import org.apache.hadoop.hive.ql.exec.UDFArgumentTypeException; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFromUnixTime.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFromUnixTime.java index 3e4351e87d..4138b4e085 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFromUnixTime.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFromUnixTime.java @@ -22,7 +22,7 @@ import java.time.ZoneId; import java.util.Date; import java.util.TimeZone; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.TimestampTZUtil; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.exec.Description; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInitCap.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInitCap.java index 164635c3a3..27dfa3e014 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInitCap.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInitCap.java @@ -20,9 +20,10 @@ import static org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils.PrimitiveGrouping.STRING_GROUP; -import org.apache.commons.lang.WordUtils; +import org.apache.commons.text.WordUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.UDFArgumentException; +import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressions; import org.apache.hadoop.hive.ql.exec.vector.expressions.StringInitCap; import org.apache.hadoop.hive.ql.metadata.HiveException; @@ -68,7 +69,7 @@ public Object evaluate(DeferredObject[] arguments) throws HiveException { return null; } - String valCap = WordUtils.capitalizeFully(val); + String valCap = WordUtils.capitalizeFully(val, Utilities.COMMON_WHITESPACE_CHARS); output.set(valCap); return output; } @@ -82,4 +83,4 @@ public String getDisplayString(String[] children) { protected String getFuncName() { return "initcap"; } -} \ No newline at end of file +} diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLTrim.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLTrim.java index bf7a3c3a30..2675b32432 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLTrim.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLTrim.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.udf.generic; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressions; import org.apache.hadoop.hive.ql.exec.vector.expressions.StringLTrim; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRTrim.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRTrim.java index bd5f3f5726..780f74ad0a 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRTrim.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFRTrim.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.udf.generic; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressions; import org.apache.hadoop.hive.ql.exec.vector.expressions.StringRTrim; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSubstringIndex.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSubstringIndex.java index a164ee1118..6a814efa0c 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSubstringIndex.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSubstringIndex.java @@ -20,7 +20,7 @@ import static org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils.PrimitiveGrouping.NUMERIC_GROUP; import static org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils.PrimitiveGrouping.STRING_GROUP; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.UDFArgumentException; import org.apache.hadoop.hive.ql.metadata.HiveException; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java index 6cb6193a63..163ad38a9f 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java @@ -23,7 +23,7 @@ import java.time.ZoneId; import java.util.TimeZone; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.TimestampTZ; import org.apache.hadoop.hive.common.type.TimestampTZUtil; import org.apache.hadoop.hive.conf.HiveConf; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrim.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrim.java index 320a9bbd85..4159c83fcf 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrim.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrim.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.udf.generic; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressions; import org.apache.hadoop.hive.ql.exec.vector.expressions.StringTrim; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java index 827e50fe63..8bf51e5b8c 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/ptf/WindowingTableFunction.java @@ -25,7 +25,7 @@ import java.util.List; import java.util.Map; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/xml/GenericUDFXPath.java ql/src/java/org/apache/hadoop/hive/ql/udf/xml/GenericUDFXPath.java index 554b3ce6e8..95f4dafa82 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/xml/GenericUDFXPath.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/xml/GenericUDFXPath.java @@ -22,7 +22,7 @@ import java.util.Collections; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.ql.exec.Description; import org.apache.hadoop.hive.ql.exec.UDFArgumentException; import org.apache.hadoop.hive.ql.metadata.HiveException; diff --git ql/src/test/org/apache/hadoop/hive/ql/exec/vector/VectorRandomRowSource.java ql/src/test/org/apache/hadoop/hive/ql/exec/vector/VectorRandomRowSource.java index ecd381d514..1b6dbe48d8 100644 --- ql/src/test/org/apache/hadoop/hive/ql/exec/vector/VectorRandomRowSource.java +++ ql/src/test/org/apache/hadoop/hive/ql/exec/vector/VectorRandomRowSource.java @@ -29,7 +29,7 @@ import java.util.Random; import java.util.Set; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.DataTypePhysicalVariation; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.common.type.HiveChar; diff --git ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorDateExpressions.java ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorDateExpressions.java index 078b2e9d34..1f81197128 100644 --- ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorDateExpressions.java +++ ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorDateExpressions.java @@ -30,7 +30,7 @@ import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory; import org.junit.Assert; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector; import org.apache.hadoop.hive.ql.exec.vector.TestVectorizedRowBatch; diff --git ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/MapJoinTestConfig.java ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/MapJoinTestConfig.java index d127342061..a250533f55 100644 --- ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/MapJoinTestConfig.java +++ ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/MapJoinTestConfig.java @@ -26,7 +26,7 @@ import java.util.Map; import java.util.Map.Entry; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.hive.common.type.DataTypePhysicalVariation; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.CompilationOpContext; diff --git ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/TestMapJoinOperator.java ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/TestMapJoinOperator.java index c4dc547c38..da74fd5991 100644 --- ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/TestMapJoinOperator.java +++ ql/src/test/org/apache/hadoop/hive/ql/exec/vector/mapjoin/TestMapJoinOperator.java @@ -21,7 +21,7 @@ import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertTrue; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.CompilationOpContext; import org.apache.hadoop.hive.ql.exec.MapJoinOperator; diff --git ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedListColumnReader.java ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedListColumnReader.java index 9aa4140a5e..03c2c7a49b 100644 --- ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedListColumnReader.java +++ ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedListColumnReader.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.io.parquet; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.type.HiveDecimal; import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; diff --git ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedMapColumnReader.java ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedMapColumnReader.java index 5b071cf7ed..c3041c7533 100644 --- ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedMapColumnReader.java +++ ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestVectorizedMapColumnReader.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.ql.io.parquet; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.common.type.HiveDecimal; import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; diff --git ql/src/test/org/apache/hadoop/hive/ql/io/parquet/VectorizedColumnReaderTestBase.java ql/src/test/org/apache/hadoop/hive/ql/io/parquet/VectorizedColumnReaderTestBase.java index 1d32afe00c..0d11381ae3 100644 --- ql/src/test/org/apache/hadoop/hive/ql/io/parquet/VectorizedColumnReaderTestBase.java +++ ql/src/test/org/apache/hadoop/hive/ql/io/parquet/VectorizedColumnReaderTestBase.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.io.parquet; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; diff --git ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java index 850db9fefb..d1415685c9 100644 --- ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java +++ ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java @@ -26,7 +26,7 @@ import java.lang.reflect.Modifier; import java.util.Set; -import org.apache.commons.lang.reflect.FieldUtils; +import org.apache.commons.lang3.reflect.FieldUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.plan.ColStatistics.Range; import org.apache.hadoop.hive.serde.serdeConstants; diff --git ql/src/test/queries/clientpositive/vector_decimal_partition.q ql/src/test/queries/clientpositive/vector_decimal_partition.q new file mode 100644 index 0000000000..dc8e667a47 --- /dev/null +++ ql/src/test/queries/clientpositive/vector_decimal_partition.q @@ -0,0 +1,16 @@ +set hive.mapred.mode=nonstrict; +set hive.explain.user=false; +SET hive.vectorized.execution.enabled=true; +set hive.fetch.task.conversion=none; + +DROP TABLE IF EXISTS decimal_part; + +CREATE TABLE decimal_part (id DECIMAL(4,0), foo VARCHAR(10)) + PARTITIONED BY (nr_bank DECIMAL(4,0)) +STORED AS ORC TBLPROPERTIES ('orc.compress'='ZLIB'); + +INSERT INTO decimal_part PARTITION (nr_bank = 88) VALUES (1, 'test'); +INSERT INTO decimal_part PARTITION (nr_bank = 8801) VALUES (1, '8801'); + +EXPLAIN VECTORIZATION EXPRESSION SELECT count(*), nr_bank FROM decimal_part GROUP BY nr_bank; +SELECT count(*), nr_bank FROM decimal_part GROUP BY nr_bank; \ No newline at end of file diff --git ql/src/test/results/clientpositive/vector_decimal_partition.q.out ql/src/test/results/clientpositive/vector_decimal_partition.q.out new file mode 100644 index 0000000000..bc85edfb7f --- /dev/null +++ ql/src/test/results/clientpositive/vector_decimal_partition.q.out @@ -0,0 +1,153 @@ +PREHOOK: query: DROP TABLE IF EXISTS decimal_part +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE IF EXISTS decimal_part +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE TABLE decimal_part (id DECIMAL(4,0), foo VARCHAR(10)) + PARTITIONED BY (nr_bank DECIMAL(4,0)) +STORED AS ORC TBLPROPERTIES ('orc.compress'='ZLIB') +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@decimal_part +POSTHOOK: query: CREATE TABLE decimal_part (id DECIMAL(4,0), foo VARCHAR(10)) + PARTITIONED BY (nr_bank DECIMAL(4,0)) +STORED AS ORC TBLPROPERTIES ('orc.compress'='ZLIB') +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@decimal_part +PREHOOK: query: INSERT INTO decimal_part PARTITION (nr_bank = 88) VALUES (1, 'test') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@decimal_part@nr_bank=88 +POSTHOOK: query: INSERT INTO decimal_part PARTITION (nr_bank = 88) VALUES (1, 'test') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@decimal_part@nr_bank=88 +POSTHOOK: Lineage: decimal_part PARTITION(nr_bank=88).foo SCRIPT [] +POSTHOOK: Lineage: decimal_part PARTITION(nr_bank=88).id SCRIPT [] +PREHOOK: query: INSERT INTO decimal_part PARTITION (nr_bank = 8801) VALUES (1, '8801') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@decimal_part@nr_bank=8801 +POSTHOOK: query: INSERT INTO decimal_part PARTITION (nr_bank = 8801) VALUES (1, '8801') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@decimal_part@nr_bank=8801 +POSTHOOK: Lineage: decimal_part PARTITION(nr_bank=8801).foo SCRIPT [] +POSTHOOK: Lineage: decimal_part PARTITION(nr_bank=8801).id SCRIPT [] +PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT count(*), nr_bank FROM decimal_part GROUP BY nr_bank +PREHOOK: type: QUERY +PREHOOK: Input: default@decimal_part +PREHOOK: Input: default@decimal_part@nr_bank=88 +PREHOOK: Input: default@decimal_part@nr_bank=8801 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT count(*), nr_bank FROM decimal_part GROUP BY nr_bank +POSTHOOK: type: QUERY +POSTHOOK: Input: default@decimal_part +POSTHOOK: Input: default@decimal_part@nr_bank=88 +POSTHOOK: Input: default@decimal_part@nr_bank=8801 +#### A masked pattern was here #### +PLAN VECTORIZATION: + enabled: true + enabledConditionsMet: [hive.vectorized.execution.enabled IS true] + +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Map Reduce + Map Operator Tree: + TableScan + alias: decimal_part + Statistics: Num rows: 2 Data size: 624 Basic stats: COMPLETE Column stats: COMPLETE + TableScan Vectorization: + native: true + Select Operator + expressions: nr_bank (type: decimal(4,0)) + outputColumnNames: nr_bank + Select Vectorization: + className: VectorSelectOperator + native: true + projectedOutputColumnNums: [2] + Statistics: Num rows: 2 Data size: 624 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + Group By Vectorization: + aggregators: VectorUDAFCountStar(*) -> bigint + className: VectorGroupByOperator + groupByMode: HASH + keyExpressions: ConvertDecimal64ToDecimal(col 2:decimal(4,0)/DECIMAL_64) -> 4:decimal(4,0) + native: false + vectorProcessingMode: HASH + projectedOutputColumnNums: [0] + keys: nr_bank (type: decimal(4,0)) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: decimal(4,0)) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: decimal(4,0)) + Reduce Sink Vectorization: + className: VectorReduceSinkOperator + native: false + nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true + nativeConditionsNotMet: hive.execution.engine mr IN [tez, spark] IS false + Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized + Map Vectorization: + enabled: true + enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true + inputFormatFeatureSupport: [DECIMAL_64] + featureSupportInUse: [DECIMAL_64] + inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + allNative: false + usesVectorUDFAdaptor: false + vectorized: true + Reduce Vectorization: + enabled: false + enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true + enableConditionsNotMet: hive.execution.engine mr IN [tez, spark] IS false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: decimal(4,0)) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: bigint), _col0 (type: decimal(4,0)) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT count(*), nr_bank FROM decimal_part GROUP BY nr_bank +PREHOOK: type: QUERY +PREHOOK: Input: default@decimal_part +PREHOOK: Input: default@decimal_part@nr_bank=88 +PREHOOK: Input: default@decimal_part@nr_bank=8801 +#### A masked pattern was here #### +POSTHOOK: query: SELECT count(*), nr_bank FROM decimal_part GROUP BY nr_bank +POSTHOOK: type: QUERY +POSTHOOK: Input: default@decimal_part +POSTHOOK: Input: default@decimal_part@nr_bank=88 +POSTHOOK: Input: default@decimal_part@nr_bank=8801 +#### A masked pattern was here #### +1 88 +1 8801 diff --git serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroLazyObjectInspector.java serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroLazyObjectInspector.java index cd37bcc7ef..5a857f2be6 100644 --- serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroLazyObjectInspector.java +++ serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroLazyObjectInspector.java @@ -29,7 +29,7 @@ import org.apache.avro.generic.GenericDatumReader; import org.apache.avro.generic.GenericRecord; import org.apache.avro.io.DatumReader; -import org.apache.commons.lang.ClassUtils; +import org.apache.commons.lang3.ClassUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.serde2.SerDeException; diff --git serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyObjectInspectorParametersImpl.java serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyObjectInspectorParametersImpl.java index f99c313b98..3172b11e10 100644 --- serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyObjectInspectorParametersImpl.java +++ serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyObjectInspectorParametersImpl.java @@ -21,8 +21,8 @@ import java.util.Arrays; import java.util.List; -import org.apache.commons.lang.ObjectUtils; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.ObjectUtils; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe; import org.apache.hadoop.io.Text; diff --git serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java index 3cb21b7c3e..9e19da7828 100644 --- serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java +++ serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataInputStream.java @@ -19,8 +19,8 @@ package org.apache.hadoop.hive.serde2.teradata; import org.apache.commons.io.input.SwappedDataInputStream; -import org.apache.commons.lang.ArrayUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.ArrayUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.type.Date; import org.apache.hadoop.hive.common.type.HiveDecimal; import org.apache.hadoop.hive.common.type.Timestamp; diff --git serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java index 3799aa21b7..a9572342fa 100644 --- serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java +++ serde/src/java/org/apache/hadoop/hive/serde2/teradata/TeradataBinaryDataOutputStream.java @@ -20,7 +20,7 @@ import org.apache.commons.io.EndianUtils; import org.apache.commons.io.output.ByteArrayOutputStream; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.serde2.io.DateWritableV2; diff --git serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java index cb48fef521..2e44074a87 100644 --- serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java +++ serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableFast.java @@ -24,7 +24,7 @@ import java.util.List; import java.util.Random; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.serde2.ByteStream.Output; import org.apache.hadoop.hive.serde2.AbstractSerDe; import org.apache.hadoop.hive.serde2.SerdeRandomRowSource; diff --git serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableSerDe.java serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableSerDe.java index e64dc8061e..77b622574c 100644 --- serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableSerDe.java +++ serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableSerDe.java @@ -21,7 +21,7 @@ import java.util.Properties; import java.util.Random; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.serde.serdeConstants; import org.apache.hadoop.hive.serde2.AbstractSerDe; diff --git serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestReflectionObjectInspectors.java serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestReflectionObjectInspectors.java index ef330543b7..e6cb07bc52 100644 --- serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestReflectionObjectInspectors.java +++ serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestReflectionObjectInspectors.java @@ -28,7 +28,7 @@ import java.util.concurrent.ScheduledExecutorService; import java.util.concurrent.TimeUnit; -import org.apache.commons.lang.mutable.MutableObject; +import org.apache.commons.lang3.mutable.MutableObject; import org.apache.commons.lang3.tuple.Pair; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector.Category; import org.apache.hadoop.hive.serde2.objectinspector.primitive.JavaConstantStringObjectInspector; diff --git service/src/java/org/apache/hive/http/JdbcJarDownloadServlet.java service/src/java/org/apache/hive/http/JdbcJarDownloadServlet.java index 142e0a192d..12f1c7b05b 100644 --- service/src/java/org/apache/hive/http/JdbcJarDownloadServlet.java +++ service/src/java/org/apache/hive/http/JdbcJarDownloadServlet.java @@ -30,7 +30,7 @@ import javax.servlet.http.HttpServletResponse; import org.apache.commons.io.filefilter.WildcardFileFilter; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; diff --git service/src/java/org/apache/hive/service/auth/LdapAuthenticationProviderImpl.java service/src/java/org/apache/hive/service/auth/LdapAuthenticationProviderImpl.java index 0120513b51..0f318fb5d8 100644 --- service/src/java/org/apache/hive/service/auth/LdapAuthenticationProviderImpl.java +++ service/src/java/org/apache/hive/service/auth/LdapAuthenticationProviderImpl.java @@ -24,7 +24,7 @@ import java.io.IOException; import java.util.Iterator; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hive.service.ServiceUtils; import org.apache.hive.service.auth.ldap.ChainFilterFactory; diff --git service/src/java/org/apache/hive/service/auth/ldap/LdapUtils.java service/src/java/org/apache/hive/service/auth/ldap/LdapUtils.java index d3caaefc6d..da1465774e 100644 --- service/src/java/org/apache/hive/service/auth/ldap/LdapUtils.java +++ service/src/java/org/apache/hive/service/auth/ldap/LdapUtils.java @@ -21,7 +21,7 @@ import java.util.Collection; import java.util.Collections; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hive.service.ServiceUtils; import org.slf4j.Logger; diff --git service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java index 81d6a7e095..475b1bb47c 100644 --- service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java +++ service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java @@ -34,7 +34,7 @@ import java.util.concurrent.Semaphore; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.cli.HiveFileProcessor; import org.apache.hadoop.hive.common.cli.IHiveFileProcessor; import org.apache.hadoop.hive.conf.HiveConf; diff --git service/src/java/org/apache/hive/service/server/KillQueryImpl.java service/src/java/org/apache/hive/service/server/KillQueryImpl.java index d9a5033b62..883e32bd2e 100644 --- service/src/java/org/apache/hive/service/server/KillQueryImpl.java +++ service/src/java/org/apache/hive/service/server/KillQueryImpl.java @@ -18,7 +18,7 @@ package org.apache.hive.service.server; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.QueryState; diff --git service/src/test/org/apache/hive/service/cli/TestHiveSQLException.java service/src/test/org/apache/hive/service/cli/TestHiveSQLException.java index 39a71b0cc4..b5e221b590 100644 --- service/src/test/org/apache/hive/service/cli/TestHiveSQLException.java +++ service/src/test/org/apache/hive/service/cli/TestHiveSQLException.java @@ -22,7 +22,7 @@ import org.junit.Assert; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hive.service.rpc.thrift.TStatus; import org.apache.hive.service.rpc.thrift.TStatusCode; import org.junit.Test; diff --git shims/0.23/src/main/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java shims/0.23/src/main/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java index f55b3ff1f3..23e7d5ef26 100644 --- shims/0.23/src/main/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java +++ shims/0.23/src/main/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java @@ -39,7 +39,7 @@ import java.util.TreeMap; import javax.security.auth.Subject; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.crypto.CipherSuite; import org.apache.hadoop.crypto.key.KeyProvider; diff --git shims/common/src/main/java/org/apache/hadoop/hive/io/HdfsUtils.java shims/common/src/main/java/org/apache/hadoop/hive/io/HdfsUtils.java index facad8af63..e59eb327f8 100644 --- shims/common/src/main/java/org/apache/hadoop/hive/io/HdfsUtils.java +++ shims/common/src/main/java/org/apache/hadoop/hive/io/HdfsUtils.java @@ -25,7 +25,7 @@ import com.google.common.annotations.VisibleForTesting; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; diff --git shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java index 8d6777c982..11506514ce 100644 --- shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java +++ shims/common/src/main/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java @@ -30,7 +30,7 @@ import java.util.HashSet; import java.util.Set; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.conf.Configuration; diff --git spark-client/src/main/java/org/apache/hive/spark/client/SparkClientUtilities.java spark-client/src/main/java/org/apache/hive/spark/client/SparkClientUtilities.java index 4f8d88046e..37aed89bb4 100644 --- spark-client/src/main/java/org/apache/hive/spark/client/SparkClientUtilities.java +++ spark-client/src/main/java/org/apache/hive/spark/client/SparkClientUtilities.java @@ -33,7 +33,7 @@ import java.util.Map; import java.util.concurrent.ConcurrentHashMap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.spark.SparkContext; import org.apache.spark.util.MutableURLClassLoader; diff --git spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcConfiguration.java spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcConfiguration.java index eb824efdfb..29cc7f1258 100644 --- spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcConfiguration.java +++ spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcConfiguration.java @@ -32,7 +32,7 @@ import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.ServerUtils; import org.apache.hadoop.hive.common.classification.InterfaceAudience; import org.apache.hadoop.hive.conf.HiveConf; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java index cc97418d16..8220f10097 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java @@ -2641,7 +2641,7 @@ public boolean deleteTableColumnStatistics(String catName, String dbName, String public List getSchema(String catName, String db, String tableName) throws TException { EnvironmentContext envCxt = null; String addedJars = MetastoreConf.getVar(conf, ConfVars.ADDED_JARS); - if (org.apache.commons.lang.StringUtils.isNotBlank(addedJars)) { + if (org.apache.commons.lang3.StringUtils.isNotBlank(addedJars)) { Map props = new HashMap<>(); props.put("hive.added.jars.path", addedJars); envCxt = new EnvironmentContext(props); diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreLdapAuthenticationProviderImpl.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreLdapAuthenticationProviderImpl.java index e9d22c5917..568758ac0e 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreLdapAuthenticationProviderImpl.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreLdapAuthenticationProviderImpl.java @@ -24,7 +24,7 @@ import java.io.IOException; import java.util.Iterator; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.ldap.ChainFilterFactory; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/Warehouse.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/Warehouse.java index 9ce379bb35..75a9368d8c 100755 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/Warehouse.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/Warehouse.java @@ -30,7 +30,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.TableName; import org.apache.hadoop.hive.metastore.api.Catalog; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/GroupFilterFactory.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/GroupFilterFactory.java index fefca20abb..8fb16ea725 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/GroupFilterFactory.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/GroupFilterFactory.java @@ -28,7 +28,7 @@ import javax.naming.NamingException; import javax.security.sasl.AuthenticationException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.slf4j.Logger; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/LdapUtils.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/LdapUtils.java index 47016a9f81..da6db19797 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/LdapUtils.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ldap/LdapUtils.java @@ -21,7 +21,7 @@ import java.util.Collection; import java.util.Collections; import java.util.List; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.ServiceUtils; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FilterUtils.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FilterUtils.java index 2ed314b4d7..d9da00dd21 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FilterUtils.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/FilterUtils.java @@ -19,7 +19,7 @@ import java.util.Collections; import java.util.List; -import static org.apache.commons.lang.StringUtils.isBlank; +import static org.apache.commons.lang3.StringUtils.isBlank; import static org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.CATALOG_DB_SEPARATOR; import org.apache.hadoop.hive.metastore.MetaStoreFilterHook; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/HdfsUtils.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/HdfsUtils.java index 21227885e3..7711d6c145 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/HdfsUtils.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/HdfsUtils.java @@ -22,7 +22,7 @@ import com.google.common.base.Objects; import com.google.common.base.Predicate; import com.google.common.collect.Iterables; -import org.apache.commons.lang.ArrayUtils; +import org.apache.commons.lang3.ArrayUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; diff --git standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java index 5bd23b7748..62f5773f9b 100644 --- standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java +++ standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java @@ -39,7 +39,7 @@ import javax.annotation.Nullable; import com.google.common.collect.Lists; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.CommonConfigurationKeysPublic; import org.apache.hadoop.fs.Path; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AuthFactory.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AuthFactory.java index 25099d3cbe..3fb6ca9256 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AuthFactory.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AuthFactory.java @@ -24,7 +24,7 @@ import javax.security.auth.login.LoginException; import javax.security.sasl.AuthenticationException; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.ConfVars; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java index 5eaf53f79b..dda407a6fc 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java @@ -22,7 +22,7 @@ import com.google.common.collect.Lists; import com.google.common.collect.Multimap; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.common.repl.ReplConst; import org.apache.hadoop.hive.common.TableName; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; @@ -1160,7 +1160,7 @@ private void checkColTypeChangeCompatible(List oldCols, List(catName, dbName, tblName, true, true) { @@ -9122,7 +9122,7 @@ public boolean deletePartitionColumnStatistics(String catName, String dbName, St throws NoSuchObjectException, MetaException, InvalidObjectException, InvalidInputException { boolean ret = false; Query query = null; - dbName = org.apache.commons.lang.StringUtils.defaultString(dbName, + dbName = org.apache.commons.lang3.StringUtils.defaultString(dbName, Warehouse.DEFAULT_DATABASE_NAME); catName = normalizeIdentifier(catName); if (tableName == null) { @@ -9225,7 +9225,7 @@ public boolean deleteTableColumnStatistics(String catName, String dbName, String throws NoSuchObjectException, MetaException, InvalidObjectException, InvalidInputException { boolean ret = false; Query query = null; - dbName = org.apache.commons.lang.StringUtils.defaultString(dbName, + dbName = org.apache.commons.lang3.StringUtils.defaultString(dbName, Warehouse.DEFAULT_DATABASE_NAME); if (tableName == null) { throw new InvalidInputException("Table name is null."); diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/PersistenceManagerProvider.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/PersistenceManagerProvider.java index 7e9522b8b7..c2e2a1cc84 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/PersistenceManagerProvider.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/PersistenceManagerProvider.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hive.metastore; import com.google.common.annotations.VisibleForTesting; +import com.google.common.base.Suppliers; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.ConfVars; @@ -88,6 +89,7 @@ private static boolean forTwoMetastoreTesting; private static int retryLimit; private static long retryInterval; + private static com.google.common.base.Supplier passwordProvider; static { Map> map = new HashMap<>(); @@ -251,7 +253,7 @@ private static void initPMF(Configuration conf) { LOG.info( "Setting MetaStore object pin classes with hive.metastore.cache.pinobjtypes=\"{}\"", objTypes); - if (org.apache.commons.lang.StringUtils.isNotEmpty(objTypes)) { + if (org.apache.commons.lang3.StringUtils.isNotEmpty(objTypes)) { String[] typeTokens = objTypes.toLowerCase().split(","); for (String type : typeTokens) { type = type.trim(); @@ -259,7 +261,7 @@ private static void initPMF(Configuration conf) { dsc.pinAll(true, PINCLASSMAP.get(type)); } else { LOG.warn("{} is not one of the pinnable object types: {}", type, - org.apache.commons.lang.StringUtils.join(PINCLASSMAP.keySet(), " ")); + org.apache.commons.lang3.StringUtils.join(PINCLASSMAP.keySet(), " ")); } } } @@ -450,16 +452,18 @@ private static Properties getDataSourceProps(Configuration conf) { */ // Password may no longer be in the conf, use getPassword() - try { - String passwd = MetastoreConf.getPassword(conf, MetastoreConf.ConfVars.PWD); - if (org.apache.commons.lang.StringUtils.isNotEmpty(passwd)) { - // We can get away with the use of varname here because varname == hiveName for PWD - prop.setProperty(ConfVars.PWD.getVarname(), passwd); + passwordProvider = passwordProvider != null ? passwordProvider : Suppliers.memoize(() -> { + try { + return MetastoreConf.getPassword(conf, ConfVars.PWD); + } catch (IOException err) { + throw new RuntimeException("Error getting metastore password: " + err.getMessage(), err); } - } catch (IOException err) { - throw new RuntimeException("Error getting metastore password: " + err.getMessage(), err); + }); + String passwd = passwordProvider.get(); + if (org.apache.commons.lang.StringUtils.isNotEmpty(passwd)) { + // We can get away with the use of varname here because varname == hiveName for PWD + prop.setProperty(ConfVars.PWD.getVarname(), passwd); } - if (LOG.isDebugEnabled()) { for (Entry e : prop.entrySet()) { if (MetastoreConf.isPrintable(e.getKey().toString())) { diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStoreProxy.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStoreProxy.java index 0add00369a..775c163e31 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStoreProxy.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RawStoreProxy.java @@ -26,7 +26,7 @@ import java.util.List; import java.util.concurrent.TimeUnit; -import org.apache.commons.lang.ClassUtils; +import org.apache.commons.lang3.ClassUtils; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java index a830eb9b0f..86945f3e4f 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java @@ -25,7 +25,7 @@ import java.lang.reflect.UndeclaredThrowableException; import java.util.concurrent.TimeUnit; -import org.apache.commons.lang.exception.ExceptionUtils; +import org.apache.commons.lang3.exception.ExceptionUtils; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CacheUtils.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CacheUtils.java index b54a11516f..c0cc0e61e8 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CacheUtils.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CacheUtils.java @@ -74,7 +74,7 @@ public static String buildTableColKey(String catName, String dbName, String tabl } public static String buildKey(String... elements) { - return org.apache.commons.lang.StringUtils.join(elements, delimit); + return org.apache.commons.lang3.StringUtils.join(elements, delimit); } public static String[] splitDbName(String key) { diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java index d36b3f014a..f99209868a 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java @@ -35,7 +35,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.lang.exception.ExceptionUtils; +import org.apache.commons.lang3.exception.ExceptionUtils; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; @@ -427,7 +427,7 @@ static void prewarm(RawStore rawStore) { Collection catalogsToCache; try { catalogsToCache = catalogsToCache(rawStore); - LOG.info("Going to cache catalogs: " + org.apache.commons.lang.StringUtils.join(catalogsToCache, ", ")); + LOG.info("Going to cache catalogs: " + org.apache.commons.lang3.StringUtils.join(catalogsToCache, ", ")); List catalogs = new ArrayList<>(catalogsToCache.size()); for (String catName : catalogsToCache) { catalogs.add(rawStore.getCatalog(catName)); diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/client/builder/ConstraintBuilder.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/client/builder/ConstraintBuilder.java index 2e32cbf3c4..c896d06e19 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/client/builder/ConstraintBuilder.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/client/builder/ConstraintBuilder.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hive.metastore.client.builder; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.MetaException; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DBTokenStore.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DBTokenStore.java index b59330cebf..f59915c14f 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DBTokenStore.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DBTokenStore.java @@ -24,7 +24,7 @@ import java.util.List; import org.apache.commons.codec.binary.Base64; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge.Server.ServerMode; import org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager.DelegationTokenInformation; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DelegationTokenTool.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DelegationTokenTool.java index 3637db09f1..44ba8d3bf7 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DelegationTokenTool.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/DelegationTokenTool.java @@ -25,7 +25,7 @@ import org.apache.commons.cli.GnuParser; import org.apache.commons.cli.Option; import org.apache.commons.cli.Options; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.fs.Path; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/ZooKeeperTokenStore.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/ZooKeeperTokenStore.java index da28fed825..af52fcc5f6 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/ZooKeeperTokenStore.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/ZooKeeperTokenStore.java @@ -24,7 +24,7 @@ import java.util.List; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.curator.framework.CuratorFramework; import org.apache.curator.framework.CuratorFrameworkFactory; import org.apache.curator.framework.api.ACLProvider; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/MetastoreSchemaTool.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/MetastoreSchemaTool.java index 223b3cb362..efc4224783 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/MetastoreSchemaTool.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/MetastoreSchemaTool.java @@ -21,7 +21,7 @@ import org.apache.commons.cli.OptionGroup; import org.apache.commons.cli.ParseException; import org.apache.commons.io.output.NullOutputStream; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.metastore.HiveMetaException; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/SchemaToolTaskValidate.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/SchemaToolTaskValidate.java index 9bf711eab9..e3428e8cc2 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/SchemaToolTaskValidate.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/SchemaToolTaskValidate.java @@ -37,8 +37,8 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.lang.ArrayUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.ArrayUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.tuple.Pair; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.metastore.HiveMetaException; diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java index 1dc3867929..29acbd1bd4 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java @@ -52,8 +52,8 @@ import javax.sql.DataSource; -import org.apache.commons.lang.ArrayUtils; -import org.apache.commons.lang.NotImplementedException; +import org.apache.commons.lang3.ArrayUtils; +import org.apache.commons.lang3.NotImplementedException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -5436,7 +5436,7 @@ public LockHandle acquireLock(String key) throws MetaException { } public void acquireLock(String key, LockHandle handle) { //the idea is that this will use LockHandle.dbConn - throw new NotImplementedException(); + throw new NotImplementedException("acquireLock(String, LockHandle) is not implemented"); } private static final class LockHandleImpl implements LockHandle { private final Connection dbConn; @@ -5458,7 +5458,7 @@ public void acquireLock(String key, LockHandle handle) { void addKey(String key) { //keys.add(key); //would need a list of (stmt,rs) pairs - 1 for each key - throw new NotImplementedException(); + throw new NotImplementedException("addKey(String) is not implemented, would require a list of (stmt,rs) pairs / key"); } @Override diff --git standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreServerUtils.java standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreServerUtils.java index a30b6bf4f7..799e879dd1 100644 --- standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreServerUtils.java +++ standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreServerUtils.java @@ -56,7 +56,7 @@ import org.apache.commons.beanutils.PropertyUtils; import org.apache.commons.collections.CollectionUtils; import org.apache.commons.collections.ListUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.Path; @@ -121,7 +121,7 @@ = new com.google.common.base.Function() { @Override public String apply(@Nullable String string) { - return org.apache.commons.lang.StringUtils.defaultString(string); + return org.apache.commons.lang3.StringUtils.defaultString(string); } }; @@ -340,7 +340,7 @@ private static String getPartitionValWithInvalidCharacter(List partVals, SortedSet sortedOuterList = new TreeSet<>(); for (List innerList : skewed.getSkewedColValues()) { SortedSet sortedInnerList = new TreeSet<>(innerList); - sortedOuterList.add(org.apache.commons.lang.StringUtils.join(sortedInnerList, ".")); + sortedOuterList.add(org.apache.commons.lang3.StringUtils.join(sortedInnerList, ".")); } for (String colval : sortedOuterList) { md.update(colval.getBytes(ENCODING)); @@ -350,7 +350,7 @@ private static String getPartitionValWithInvalidCharacter(List partVals, SortedMap sortedMap = new TreeMap<>(); for (Map.Entry, String> smap : skewed.getSkewedColValueLocationMaps().entrySet()) { SortedSet sortedKey = new TreeSet<>(smap.getKey()); - sortedMap.put(org.apache.commons.lang.StringUtils.join(sortedKey, "."), smap.getValue()); + sortedMap.put(org.apache.commons.lang3.StringUtils.join(sortedKey, "."), smap.getValue()); } for (Map.Entry e : sortedMap.entrySet()) { md.update(e.getKey().getBytes(ENCODING)); diff --git standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java index 1faf5733a7..fc071f9a20 100644 --- standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java +++ standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java @@ -1852,7 +1852,7 @@ public boolean deleteTableColumnStatistics(String dbName, String tableName, Stri UnknownDBException { EnvironmentContext envCxt = null; String addedJars = MetastoreConf.getVar(conf, ConfVars.ADDED_JARS); - if(org.apache.commons.lang.StringUtils.isNotBlank(addedJars)) { + if(org.apache.commons.lang3.StringUtils.isNotBlank(addedJars)) { Map props = new HashMap(); props.put("hive.added.jars.path", addedJars); envCxt = new EnvironmentContext(props); diff --git standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/VerifyingObjectStore.java standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/VerifyingObjectStore.java index c5cbeb34f0..1da04c760a 100644 --- standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/VerifyingObjectStore.java +++ standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/VerifyingObjectStore.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.metastore; -import static org.apache.commons.lang.StringUtils.repeat; +import static org.apache.commons.lang3.StringUtils.repeat; import static org.apache.hadoop.hive.metastore.Warehouse.DEFAULT_CATALOG_NAME; import java.lang.reflect.AccessibleObject; @@ -32,8 +32,8 @@ import java.util.Map; import java.util.Set; -import org.apache.commons.lang.ClassUtils; -import org.apache.commons.lang.builder.EqualsBuilder; +import org.apache.commons.lang3.ClassUtils; +import org.apache.commons.lang3.builder.EqualsBuilder; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.metastore.api.ColumnStatistics; diff --git standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java index e53720d173..4ab06c7fc9 100644 --- standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java +++ standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java @@ -25,7 +25,7 @@ import java.util.Map; import java.util.stream.Collectors; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.fs.Path; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.IMetaStoreClient; diff --git standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/dbinstall/rules/DatabaseRule.java standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/dbinstall/rules/DatabaseRule.java index c1f49d8d46..3f82891ef6 100644 --- standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/dbinstall/rules/DatabaseRule.java +++ standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/dbinstall/rules/DatabaseRule.java @@ -25,7 +25,7 @@ import java.util.List; import java.util.concurrent.TimeUnit; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.hive.metastore.tools.schematool.MetastoreSchemaTool; import org.junit.rules.ExternalResource; import org.slf4j.Logger; diff --git standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/tools/schematool/TestSchemaToolForMetastore.java standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/tools/schematool/TestSchemaToolForMetastore.java index 1f7d17f03f..0d3446ce65 100644 --- standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/tools/schematool/TestSchemaToolForMetastore.java +++ standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/tools/schematool/TestSchemaToolForMetastore.java @@ -33,7 +33,7 @@ import org.apache.commons.dbcp.DelegatingConnection; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.text.StrTokenizer; +import org.apache.commons.lang3.text.StrTokenizer; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.HiveMetaException; import org.apache.hadoop.hive.metastore.IMetaStoreSchemaInfo; diff --git storage-api/src/java/org/apache/hadoop/hive/common/io/CacheTag.java storage-api/src/java/org/apache/hadoop/hive/common/io/CacheTag.java index 3062caa557..ba6e5344ea 100644 --- storage-api/src/java/org/apache/hadoop/hive/common/io/CacheTag.java +++ storage-api/src/java/org/apache/hadoop/hive/common/io/CacheTag.java @@ -23,7 +23,7 @@ import java.util.HashMap; import java.util.Map; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; /** * Used for identifying the related object of the buffer stored in cache. diff --git storage-api/src/java/org/apache/hadoop/hive/common/type/HiveIntervalDayTime.java storage-api/src/java/org/apache/hadoop/hive/common/type/HiveIntervalDayTime.java index 6723b7d641..907f6607a6 100644 --- storage-api/src/java/org/apache/hadoop/hive/common/type/HiveIntervalDayTime.java +++ storage-api/src/java/org/apache/hadoop/hive/common/type/HiveIntervalDayTime.java @@ -22,7 +22,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.lang.builder.HashCodeBuilder; +import org.apache.commons.lang3.builder.HashCodeBuilder; import org.apache.hive.common.util.IntervalDayTimeUtils; diff --git testutils/ptest2/pom.xml testutils/ptest2/pom.xml index c7bf6eb670..6d4305621b 100644 --- testutils/ptest2/pom.xml +++ testutils/ptest2/pom.xml @@ -60,9 +60,9 @@ limitations under the License. 1.2 - commons-lang - commons-lang - 2.6 + org.apache.commons + commons-lang3 + 3.9 com.google.guava diff --git testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/HostExecutor.java testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/HostExecutor.java index 47347ebfd4..8982afd75b 100644 --- testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/HostExecutor.java +++ testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/HostExecutor.java @@ -31,7 +31,7 @@ import java.util.concurrent.atomic.AtomicLong; import com.google.common.base.Stopwatch; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.apache.hive.ptest.execution.conf.Host; import org.apache.hive.ptest.execution.conf.TestBatch; import org.apache.hive.ptest.execution.ssh.RSyncCommand; diff --git testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/conf/UnitTestPropertiesParser.java testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/conf/UnitTestPropertiesParser.java index a482fce8ed..490c23bf1a 100644 --- testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/conf/UnitTestPropertiesParser.java +++ testutils/ptest2/src/main/java/org/apache/hive/ptest/execution/conf/UnitTestPropertiesParser.java @@ -40,7 +40,7 @@ import com.google.common.collect.Lists; import com.google.common.collect.Sets; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.StringUtils; +import org.apache.commons.lang3.StringUtils; import org.slf4j.Logger; class UnitTestPropertiesParser {