commit b535fb22ce494ea3f5a29e38f92c5a673cacf5e9 Author: Owen O'Malley Date: Wed Jun 1 10:58:34 2016 -0700 HIVE-13906. Remove guava from compile time dependence in the storage-api module. diff --git storage-api/pom.xml storage-api/pom.xml index 4095ed2..4767be6 100644 --- storage-api/pom.xml +++ storage-api/pom.xml @@ -32,14 +32,23 @@ - + + + commons-lang + commons-lang + ${commons-lang.version} + org.apache.hadoop hadoop-common ${hadoop.version} - true + provided + com.google.guava + guava + + commmons-logging commons-logging @@ -52,7 +61,20 @@ slf4j-log4j12 - + + + org.slf4j + slf4j-api + ${slf4j.version} + + + + + com.google.guava + guava + ${guava.version} + test + junit junit diff --git storage-api/src/java/org/apache/hadoop/hive/common/DiskRangeInfo.java storage-api/src/java/org/apache/hadoop/hive/common/DiskRangeInfo.java index 86b838c..e5025bf 100644 --- storage-api/src/java/org/apache/hadoop/hive/common/DiskRangeInfo.java +++ storage-api/src/java/org/apache/hadoop/hive/common/DiskRangeInfo.java @@ -17,12 +17,11 @@ */ package org.apache.hadoop.hive.common; +import java.util.ArrayList; import java.util.List; import org.apache.hadoop.hive.common.io.DiskRange; -import com.google.common.collect.Lists; - /** * Disk range information class containing disk ranges and total length. */ @@ -31,7 +30,7 @@ long totalLength; public DiskRangeInfo(int indexBaseOffset) { - this.diskRanges = Lists.newArrayList(); + this.diskRanges = new ArrayList<>(); // Some data is missing from the stream for PPD uncompressed read (because index offset is // relative to the entire stream and we only read part of stream if RGs are filtered; unlike // with compressed data where PPD only filters CBs, so we always get full CB, and index offset diff --git storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java index 33b70c2..4a745e4 100644 --- storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java +++ storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java @@ -18,8 +18,6 @@ package org.apache.hadoop.hive.ql.util; -import com.google.common.annotations.VisibleForTesting; - import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -257,7 +255,7 @@ public static int alignUp(int value, int align) { private static final JavaDataModel MODEL_FOR_SYSTEM = getModelForSystem(); } - @VisibleForTesting + //@VisibleForTesting static JavaDataModel getModelForSystem() { String props = null; try { diff --git storage-api/src/java/org/apache/hive/common/util/BloomFilter.java storage-api/src/java/org/apache/hive/common/util/BloomFilter.java index bb0b8f2..e60690d 100644 --- storage-api/src/java/org/apache/hive/common/util/BloomFilter.java +++ storage-api/src/java/org/apache/hive/common/util/BloomFilter.java @@ -21,8 +21,6 @@ import java.util.Arrays; import java.util.List; -import static com.google.common.base.Preconditions.checkArgument; - /** * BloomFilter is a probabilistic data structure for set membership check. BloomFilters are * highly space efficient when compared to using a HashSet. Because of the probabilistic nature of @@ -54,6 +52,12 @@ public BloomFilter(long expectedEntries) { this(expectedEntries, DEFAULT_FPP); } + static void checkArgument(boolean expression, String message) { + if (!expression) { + throw new IllegalArgumentException(message); + } + } + public BloomFilter(long expectedEntries, double fpp) { checkArgument(expectedEntries > 0, "expectedEntries should be > 0"); checkArgument(fpp > 0.0 && fpp < 1.0, "False positive probability should be > 0.0 & < 1.0");