From 0128fed9dd59e5197413780a298feb451e217cc0 Mon Sep 17 00:00:00 2001 From: "subrat.mishra" Date: Mon, 15 Oct 2018 18:33:48 +0530 Subject: [PATCH] HBASE-21263 Mention compression algorithm along with other storefile details Signed-off-by: Andrew Purtell Amending-Author: Andrew Purtell Conflicts: hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java --- .../org/apache/hadoop/hbase/regionserver/HStore.java | 7 +++++++ .../hadoop/hbase/regionserver/compactions/Compactor.java | 1 + .../org/apache/hadoop/hbase/io/hfile/TestChecksum.java | 9 +++++---- .../org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java | 7 ++++--- .../hadoop/hbase/io/hfile/TestHFileBlockIndex.java | 2 +- .../hadoop/hbase/regionserver/CreateRandomStoreFile.java | 2 +- 6 files changed, 19 insertions(+), 9 deletions(-) diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java index 719f234dae..75c38a6cd1 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java @@ -302,6 +302,7 @@ public class HStore implements Store { "hbase.hstore.flush.retries.number must be > 0, not " + flushRetriesNumber); } + // Crypto context for new store files String cipherName = family.getEncryptionType(); if (cipherName != null) { @@ -358,6 +359,12 @@ public class HStore implements Store { cryptoContext.setCipher(cipher); cryptoContext.setKey(key); } + + LOG.info("Store=" + getColumnFamilyName() + + ", memstore type=" + this.memstore.getClass().getSimpleName() + + ", storagePolicy=" + policyName + ", verifyBulkLoads=" + verifyBulkLoads + + ", encoding=" + family.getDataBlockEncoding() + + ", compression=" + family.getCompressionType()); } /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java index af8ffe81ac..b7d27de6a7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java @@ -199,6 +199,7 @@ public abstract class Compactor { ", bloomtype=" + r.getBloomFilterType().toString() + ", size=" + TraditionalBinaryPrefix.long2String(r.length(), "", 1) + ", encoding=" + r.getHFileReader().getDataBlockEncoding() + + ", compression=" + compactionCompression + ", seqNum=" + seqNum + (allFiles ? ", earliestPutTs=" + earliestPutTs: "")); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java index 99964d7309..596c17a161 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java @@ -295,10 +295,11 @@ public class TestChecksum { dataSize + HConstants.HFILEBLOCK_HEADER_SIZE, bytesPerChecksum); LOG.info("testChecksumChunks: pread=" + pread + - ", bytesPerChecksum=" + bytesPerChecksum + - ", fileSize=" + totalSize + - ", dataSize=" + dataSize + - ", expectedChunks=" + expectedChunks); + ", bytesPerChecksum=" + bytesPerChecksum + + ", fileSize=" + totalSize + + ", dataSize=" + dataSize + + ", expectedChunks=" + expectedChunks + + ", compression=" + algo); // Verify hbase checksums. assertEquals(true, hfs.useHBaseChecksum()); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java index da54e938c3..a67b66ebae 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java @@ -364,6 +364,8 @@ public class TestHFileBlock { for (Compression.Algorithm algo : COMPRESSION_ALGORITHMS) { for (boolean pread : new boolean[] { false, true }) { for (DataBlockEncoding encoding : DataBlockEncoding.values()) { + LOG.info("testDataBlockEncoding: Compression algorithm=" + algo + ", pread=" + pread + + ", dataBlockEncoder=" + encoding); Path path = new Path(TEST_UTIL.getDataTestDir(), "blocks_v2_" + algo + "_" + encoding.toString()); FSDataOutputStream os = fs.create(path); @@ -520,9 +522,8 @@ public class TestHFileBlock { for (boolean pread : BOOLEAN_VALUES) { for (boolean cacheOnWrite : BOOLEAN_VALUES) { Random rand = defaultRandom(); - LOG.info("testPreviousOffset:Compression algorithm: " + algo + - ", pread=" + pread + - ", cacheOnWrite=" + cacheOnWrite); + LOG.info("testPreviousOffset: Compression algorithm=" + algo + ", pread=" + pread + + ", cacheOnWrite=" + cacheOnWrite); Path path = new Path(TEST_UTIL.getDataTestDir(), "prev_offset"); List expectedOffsets = new ArrayList(); List expectedPrevOffsets = new ArrayList(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java index c677ccfaf4..5be8194e6d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java @@ -188,7 +188,7 @@ public class TestHFileBlockIndex { private void readIndex(boolean useTags) throws IOException { long fileSize = fs.getFileStatus(path).getLen(); - LOG.info("Size of " + path + ": " + fileSize); + LOG.info("Size of " + path + ": " + fileSize + ", compression=" + compr); FSDataInputStream istream = fs.open(path); HFileContext meta = new HFileContextBuilder() diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java index c3f7b82080..bcba1f9e99 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java @@ -209,7 +209,7 @@ public class CreateRandomStoreFile { Path storeFilePath = sfw.getPath(); long fileSize = fs.getFileStatus(storeFilePath).getLen(); - LOG.info("Created " + storeFilePath + ", " + fileSize + " bytes"); + LOG.info("Created " + storeFilePath + ", " + fileSize + " bytes, compression=" + compr); return true; } -- 2.19.0