From 25b5322bffcefc211b267dd4f2370258d3f8ff8c Mon Sep 17 00:00:00 2001 From: Viraj Jasani Date: Wed, 20 Nov 2019 22:44:57 +0530 Subject: [PATCH] HBASE-23279 : ROW_INDEX_V1 as default DataBlock encoding --- .../org/apache/hadoop/hbase/HColumnDescriptor.java | 3 ++- .../client/ColumnFamilyDescriptorBuilder.java | 14 ++++++-------- .../hbase/io/encoding/DataBlockEncoding.java | 3 +++ .../apache/hadoop/hbase/io/hfile/HFileContext.java | 2 +- .../hadoop/hbase/io/hfile/HFileContextBuilder.java | 2 +- .../hbase/io/hfile/HFileDataBlockEncoderImpl.java | 4 ++-- .../hbase/io/hfile/bucket/BucketAllocator.java | 1 + .../hadoop/hbase/io/hfile/TestHFileWriterV3.java | 4 +++- 8 files changed, 19 insertions(+), 14 deletions(-) diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java index c3919df3ab..2bbe10fa91 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -80,7 +80,8 @@ public class HColumnDescriptor implements ColumnFamilyDescriptor, Comparable DataBlockEncoding.valueOf(n.toUpperCase()), DataBlockEncoding.NONE); + n -> DataBlockEncoding.valueOf(n.toUpperCase()), DataBlockEncoding + .DEFAULT_DATA_BLOCK_ENCODING); } /** @@ -807,7 +804,8 @@ public class ColumnFamilyDescriptorBuilder { * @return this (for chained invocation) */ public ModifyableColumnFamilyDescriptor setDataBlockEncoding(DataBlockEncoding type) { - return setValue(DATA_BLOCK_ENCODING_BYTES, type == null ? DataBlockEncoding.NONE.name() : type.name()); + return setValue(DATA_BLOCK_ENCODING_BYTES, type == null ? + DataBlockEncoding.DEFAULT_DATA_BLOCK_ENCODING.name() : type.name()); } /** diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java index 335488b075..8a19b8d406 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java @@ -79,6 +79,9 @@ public enum DataBlockEncoding { this.encoderCls = encoderClsName; } + public static final DataBlockEncoding DEFAULT_DATA_BLOCK_ENCODING = + DataBlockEncoding.ROW_INDEX_V1; + /** * @return name converted to bytes. */ diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java index d606497e2f..0f0fa9c1e0 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContext.java @@ -60,7 +60,7 @@ public class HFileContext implements HeapSize, Cloneable { private int bytesPerChecksum = DEFAULT_BYTES_PER_CHECKSUM; /** Number of uncompressed bytes we allow per block. */ private int blocksize = HConstants.DEFAULT_BLOCKSIZE; - private DataBlockEncoding encoding = DataBlockEncoding.NONE; + private DataBlockEncoding encoding = DataBlockEncoding.DEFAULT_DATA_BLOCK_ENCODING; /** Encryption algorithm and key used */ private Encryption.Context cryptoContext = Encryption.Context.NONE; private long fileCreateTime; diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContextBuilder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContextBuilder.java index 5fa56264f3..ed1134bd99 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContextBuilder.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileContextBuilder.java @@ -48,7 +48,7 @@ public class HFileContextBuilder { private int bytesPerChecksum = DEFAULT_BYTES_PER_CHECKSUM; /** Number of uncompressed bytes we allow per block. */ private int blocksize = HConstants.DEFAULT_BLOCKSIZE; - private DataBlockEncoding encoding = DataBlockEncoding.NONE; + private DataBlockEncoding encoding = DataBlockEncoding.DEFAULT_DATA_BLOCK_ENCODING; /** Crypto context */ private Encryption.Context cryptoContext = Encryption.Context.NONE; private long fileCreateTime = 0; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java index 347b1f3c59..8fa9f86514 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java @@ -42,7 +42,7 @@ public class HFileDataBlockEncoderImpl implements HFileDataBlockEncoder { * @param encoding What kind of data block encoding will be used. */ public HFileDataBlockEncoderImpl(DataBlockEncoding encoding) { - this.encoding = encoding != null ? encoding : DataBlockEncoding.NONE; + this.encoding = encoding != null ? encoding : DataBlockEncoding.DEFAULT_DATA_BLOCK_ENCODING; } public static HFileDataBlockEncoder createFromFileInfo( @@ -129,7 +129,7 @@ public class HFileDataBlockEncoderImpl implements HFileDataBlockEncoder { @Override public void startBlockEncoding(HFileBlockEncodingContext encodingCtx, DataOutputStream out) throws IOException { - if (this.encoding != null && this.encoding != DataBlockEncoding.NONE) { + if (this.encoding != DataBlockEncoding.NONE) { this.encoding.getEncoder().startBlockEncoding(encodingCtx, out); } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketAllocator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketAllocator.java index 2883ff2f16..7784a2184d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketAllocator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketAllocator.java @@ -305,6 +305,7 @@ public final class BucketAllocator { BucketAllocator(long availableSpace, int[] bucketSizes) throws BucketAllocatorException { this.bucketSizes = bucketSizes == null ? DEFAULT_BUCKET_SIZES : bucketSizes; + LOG.info("BucketSizes during Bucket Allocation: {}", this.bucketSizes); Arrays.sort(this.bucketSizes); this.bigItemSize = Ints.max(this.bucketSizes); this.bucketCapacity = FEWEST_ITEMS_IN_BUCKET * (long) bigItemSize; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileWriterV3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileWriterV3.java index 4f4d36b2bc..44789669d8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileWriterV3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileWriterV3.java @@ -47,6 +47,7 @@ import org.apache.hadoop.hbase.io.ByteBuffAllocator; import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper; import org.apache.hadoop.hbase.io.compress.Compression; import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; +import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.nio.ByteBuff; import org.apache.hadoop.hbase.testclassification.IOTests; import org.apache.hadoop.hbase.testclassification.SmallTests; @@ -127,7 +128,8 @@ public class TestHFileWriterV3 { HFileContext context = new HFileContextBuilder() .withBlockSize(4096) .withIncludesTags(useTags) - .withCompression(compressAlgo).build(); + .withCompression(compressAlgo) + .withDataBlockEncoding(DataBlockEncoding.NONE).build(); CacheConfig cacheConfig = new CacheConfig(conf); HFile.Writer writer = new HFile.WriterFactory(conf, cacheConfig) .withPath(fs, hfilePath) -- 2.17.2 (Apple Git-113)