diff --git common/src/java/org/apache/hadoop/hive/common/JvmMetricsInfo.java common/src/java/org/apache/hadoop/hive/common/JvmMetricsInfo.java index 3ab73c5..76dca04 100644 --- common/src/java/org/apache/hadoop/hive/common/JvmMetricsInfo.java +++ common/src/java/org/apache/hadoop/hive/common/JvmMetricsInfo.java @@ -18,6 +18,7 @@ package org.apache.hadoop.hive.common; +import com.google.common.base.MoreObjects; import com.google.common.base.Objects; import org.apache.hadoop.metrics2.MetricsInfo; @@ -58,8 +59,8 @@ @Override public String description() { return desc; } @Override public String toString() { - return Objects.toStringHelper(this) - .add("name", name()).add("description", desc) + return MoreObjects.toStringHelper(this) + .add("name", name()).add("description", desc) .toString(); } } diff --git common/src/java/org/apache/hadoop/hive/common/JvmPauseMonitor.java common/src/java/org/apache/hadoop/hive/common/JvmPauseMonitor.java index cf080e3..fb7d7de 100644 --- common/src/java/org/apache/hadoop/hive/common/JvmPauseMonitor.java +++ common/src/java/org/apache/hadoop/hive/common/JvmPauseMonitor.java @@ -173,7 +173,7 @@ public String toString() { private class Monitor implements Runnable { @Override public void run() { - Stopwatch sw = new Stopwatch(); + Stopwatch sw = Stopwatch.createUnstarted(); Map gcTimesBeforeSleep = getGcTimes(); while (shouldRun) { sw.reset().start(); diff --git druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidSelectQueryRecordReader.java druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidSelectQueryRecordReader.java index c0744b5..0cef3a0 100644 --- druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidSelectQueryRecordReader.java +++ druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidSelectQueryRecordReader.java @@ -19,6 +19,7 @@ import java.io.IOException; import java.io.InputStream; +import java.util.Collections; import java.util.Iterator; import java.util.List; @@ -48,7 +49,7 @@ private Result current; - private Iterator values = Iterators.emptyIterator(); + private Iterator values = Collections.emptyIterator(); @Override protected JavaType getResultTypeDef() { diff --git druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidTopNQueryRecordReader.java druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidTopNQueryRecordReader.java index 6e1fffe..b39e799 100644 --- druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidTopNQueryRecordReader.java +++ druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidTopNQueryRecordReader.java @@ -19,6 +19,7 @@ import java.io.IOException; import java.io.InputStream; +import java.util.Collections; import java.util.Iterator; import java.util.List; @@ -46,7 +47,7 @@ private Result current; - private Iterator values = Iterators.emptyIterator(); + private Iterator values = Collections.emptyIterator(); @Override protected JavaType getResultTypeDef() { diff --git druid-handler/src/test/org/apache/hadoop/hive/ql/io/TestDruidRecordWriter.java druid-handler/src/test/org/apache/hadoop/hive/ql/io/TestDruidRecordWriter.java index af75bfb..8d09d27 100644 --- druid-handler/src/test/org/apache/hadoop/hive/ql/io/TestDruidRecordWriter.java +++ druid-handler/src/test/org/apache/hadoop/hive/ql/io/TestDruidRecordWriter.java @@ -64,7 +64,6 @@ import org.joda.time.DateTime; import org.joda.time.Interval; import org.junit.Assert; -import org.junit.Ignore; import org.junit.Rule; import org.junit.Test; import org.junit.rules.TemporaryFolder; @@ -116,8 +115,7 @@ public void testTimeStampColumnName() { DruidStorageHandlerUtils.DEFAULT_TIMESTAMP_COLUMN, DruidTable.DEFAULT_TIMESTAMP_COLUMN ); } - // This test fails due to conflict of guava classes with hive-exec jar. - @Ignore + @Test public void testWrite() throws IOException, SegmentLoadingException { diff --git itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCliDriver.java itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCliDriver.java index a1762ec..d97ab3b 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCliDriver.java +++ itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCliDriver.java @@ -40,7 +40,7 @@ private static final Logger LOG = LoggerFactory.getLogger(CoreCliDriver.class); private static QTestUtil qt; - + public CoreCliDriver(AbstractCliConfig testCliConfig) { super(testCliConfig); } @@ -153,7 +153,7 @@ public Void invokeInternal() throws Exception { @Override public void runTest(String tname, String fname, String fpath) throws Exception { - Stopwatch sw = new Stopwatch().start(); + Stopwatch sw = Stopwatch.createStarted(); boolean skipped = false; boolean failed = false; try { diff --git itests/util/src/main/java/org/apache/hadoop/hive/util/ElapsedTimeLoggingWrapper.java itests/util/src/main/java/org/apache/hadoop/hive/util/ElapsedTimeLoggingWrapper.java index 061a918..74f50ba 100644 --- itests/util/src/main/java/org/apache/hadoop/hive/util/ElapsedTimeLoggingWrapper.java +++ itests/util/src/main/java/org/apache/hadoop/hive/util/ElapsedTimeLoggingWrapper.java @@ -28,7 +28,7 @@ public abstract T invokeInternal() throws Exception; public T invoke(String message, Logger LOG, boolean toStdErr) throws Exception { - Stopwatch sw = new Stopwatch().start(); + Stopwatch sw = Stopwatch.createStarted(); try { T retVal = invokeInternal(); return retVal; diff --git llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskRunnerCallable.java llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskRunnerCallable.java index 87a692f..7971eda 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskRunnerCallable.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskRunnerCallable.java @@ -114,8 +114,8 @@ private final String queryId; private final HadoopShim tezHadoopShim; private boolean shouldRunTask = true; - final Stopwatch runtimeWatch = new Stopwatch(); - final Stopwatch killtimerWatch = new Stopwatch(); + final Stopwatch runtimeWatch = Stopwatch.createUnstarted(); + final Stopwatch killtimerWatch = Stopwatch.createUnstarted(); private final AtomicBoolean isStarted = new AtomicBoolean(false); private final AtomicBoolean isCompleted = new AtomicBoolean(false); private final AtomicBoolean killInvoked = new AtomicBoolean(false); @@ -290,7 +290,7 @@ public LlapTaskUmbilicalProtocol run() throws Exception { FileSystem.closeAllForUGI(fsTaskUgi); fragmentInfo.getQueryInfo().returnUmbilicalUgi(taskOwner); LOG.info("ExecutionTime for Container: " + request.getContainerIdString() + "=" + - runtimeWatch.stop().elapsedMillis()); + runtimeWatch.stop().elapsed(TimeUnit.MILLISECONDS)); if (LOG.isDebugEnabled()) { LOG.debug( "canFinish post completion: " + taskSpec.getTaskAttemptID() + ": " + canFinish()); @@ -517,14 +517,14 @@ public void onSuccess(TaskRunner2Result result) { LOG.info("Killed task {}", requestId); if (killtimerWatch.isRunning()) { killtimerWatch.stop(); - long elapsed = killtimerWatch.elapsedMillis(); + long elapsed = killtimerWatch.elapsed(TimeUnit.MILLISECONDS); LOG.info("Time to die for task {}", elapsed); if (metrics != null) { metrics.addMetricsPreemptionTimeToKill(elapsed); } } if (metrics != null) { - metrics.addMetricsPreemptionTimeLost(runtimeWatch.elapsedMillis()); + metrics.addMetricsPreemptionTimeLost(runtimeWatch.elapsed(TimeUnit.MILLISECONDS)); metrics.incrExecutorTotalKilled(); } break; diff --git llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonCacheInfo.java llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonCacheInfo.java index 427a0b1..9b302d5 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonCacheInfo.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonCacheInfo.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hive.llap.metrics; +import com.google.common.base.MoreObjects; import org.apache.hadoop.metrics2.MetricsInfo; import com.google.common.base.Objects; @@ -50,7 +51,7 @@ public String description() { @Override public String toString() { - return Objects.toStringHelper(this) + return MoreObjects.toStringHelper(this) .add("name", name()).add("description", desc) .toString(); } diff --git llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonExecutorInfo.java llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonExecutorInfo.java index 69d1c6f..63ce2be 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonExecutorInfo.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonExecutorInfo.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hive.llap.metrics; +import com.google.common.base.MoreObjects; import org.apache.hadoop.metrics2.MetricsInfo; import com.google.common.base.Objects; @@ -74,7 +75,7 @@ public String description() { @Override public String toString() { - return Objects.toStringHelper(this) + return MoreObjects.toStringHelper(this) .add("name", name()).add("description", desc) .toString(); } diff --git llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonIOInfo.java llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonIOInfo.java index f0fde62..979b2cc 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonIOInfo.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonIOInfo.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hive.llap.metrics; +import com.google.common.base.MoreObjects; import org.apache.hadoop.metrics2.MetricsInfo; import com.google.common.base.Objects; @@ -42,7 +43,7 @@ public String description() { @Override public String toString() { - return Objects.toStringHelper(this) + return MoreObjects.toStringHelper(this) .add("name", name()).add("description", desc) .toString(); } diff --git llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonJvmInfo.java llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonJvmInfo.java index a817ed0..16e94c4 100644 --- llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonJvmInfo.java +++ llap-server/src/java/org/apache/hadoop/hive/llap/metrics/LlapDaemonJvmInfo.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hive.llap.metrics; +import com.google.common.base.MoreObjects; import org.apache.hadoop.metrics2.MetricsInfo; import com.google.common.base.Objects; @@ -56,7 +57,7 @@ public String description() { @Override public String toString() { - return Objects.toStringHelper(this) + return MoreObjects.toStringHelper(this) .add("name", name()).add("description", desc) .toString(); } diff --git llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapTaskSchedulerInfo.java llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapTaskSchedulerInfo.java index c190be8..014426f 100644 --- llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapTaskSchedulerInfo.java +++ llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/metrics/LlapTaskSchedulerInfo.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hive.llap.tezplugins.metrics; +import com.google.common.base.MoreObjects; import org.apache.hadoop.metrics2.MetricsInfo; import com.google.common.base.Objects; @@ -52,7 +53,7 @@ public String description() { @Override public String toString() { - return Objects.toStringHelper(this) + return MoreObjects.toStringHelper(this) .add("name", name()).add("description", desc) .toString(); } diff --git pom.xml pom.xml index 1682f47..0c898ff 100644 --- pom.xml +++ pom.xml @@ -141,7 +141,7 @@ 3.1.0 0.1.2 0.10.1 - 14.0.1 + 19.0 2.4.11 1.3.166 3.0.0-beta1 diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java index b85a243..9d380ac 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java @@ -22,6 +22,7 @@ import java.io.Serializable; import java.util.ArrayList; import java.util.Arrays; +import java.util.Collections; import java.util.Comparator; import java.util.HashMap; import java.util.Iterator; @@ -110,7 +111,7 @@ private transient Iterator iterPath; private transient Iterator iterPartDesc; - private transient Iterator iterSplits = Iterators.emptyIterator(); + private transient Iterator iterSplits = Collections.emptyIterator(); private transient Path currPath; private transient PartitionDesc currDesc; @@ -592,7 +593,7 @@ public void clearFetchContext() throws HiveException { this.currPath = null; this.iterPath = null; this.iterPartDesc = null; - this.iterSplits = Iterators.emptyIterator(); + this.iterSplits = Collections.emptyIterator(); } catch (Exception e) { throw new HiveException("Failed with exception " + e.getMessage() + StringUtils.stringifyException(e)); diff --git ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java index 2f764f8..73db8b9 100644 --- ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java +++ ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java @@ -47,6 +47,7 @@ import org.slf4j.LoggerFactory; import java.io.IOException; +import java.nio.charset.Charset; import java.util.ArrayList; import java.util.Collection; import java.util.HashSet; @@ -463,7 +464,7 @@ private void writeVertices(JsonWriter writer, Set vertices) throws IOExc */ private String getQueryHash(String queryStr) { Hasher hasher = Hashing.md5().newHasher(); - hasher.putString(queryStr); + hasher.putBytes(queryStr.getBytes(Charset.defaultCharset())); return hasher.hash().toString(); } } diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveMetaStoreChecker.java ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveMetaStoreChecker.java index 5737c66..fe32bf0 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveMetaStoreChecker.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveMetaStoreChecker.java @@ -432,7 +432,7 @@ private void checkPartitionDirs(Path basePath, Set allDirs, final List aggPullUpPredicates = new ArrayList<>(); - final RexBuilder rexBuilder = agg.getCluster().getRexBuilder(); + final RexBuilder rexBuilder = agg.getCluster().getRexBuilder(); ImmutableBitSet groupKeys = agg.getGroupSet(); Mapping m = Mappings.create(MappingType.PARTIAL_FUNCTION, @@ -480,10 +481,10 @@ public RelOptPredicateList inferPredicates( } return RelOptPredicateList.of(rexBuilder, pulledUpPredicates, leftInferredPredicates, rightInferredPredicates); - case LEFT: - return RelOptPredicateList.of(rexBuilder, + case LEFT: + return RelOptPredicateList.of(rexBuilder, leftPreds, EMPTY_LIST, rightInferredPredicates); - case RIGHT: + case RIGHT: return RelOptPredicateList.of(rexBuilder, rightPreds, leftInferredPredicates, EMPTY_LIST); default: @@ -534,7 +535,7 @@ private void infer(List predicates, Set allExprsDigests, public Iterator iterator() { ImmutableBitSet fields = exprFields.get(predicate.toString()); if (fields.cardinality() == 0) { - return Iterators.emptyIterator(); + return Collections.emptyIterator(); } return new ExprsItr(fields); }