diff --git src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java index 10700c1..ec61c3a 100644 --- src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java +++ src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java @@ -21,14 +21,9 @@ import java.io.DataOutput; import java.io.IOException; import java.io.OutputStream; -import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.hadoop.classification.InterfaceStability; - /** * OutputStream implementation that wraps a DataOutput. */ -@InterfaceAudience.Private -@InterfaceStability.Unstable class DataOutputOutputStream extends OutputStream { private final DataOutput out; diff --git src/main/java/org/apache/hadoop/hbase/mapreduce/HLogInputFormat.java src/main/java/org/apache/hadoop/hbase/mapreduce/HLogInputFormat.java index 747063c..bdac1f4 100644 --- src/main/java/org/apache/hadoop/hbase/mapreduce/HLogInputFormat.java +++ src/main/java/org/apache/hadoop/hbase/mapreduce/HLogInputFormat.java @@ -27,7 +27,6 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -45,7 +44,6 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext; /** * Simple {@link InputFormat} for {@link HLog} files. */ -@InterfaceAudience.Public public class HLogInputFormat extends InputFormat { private static Log LOG = LogFactory.getLog(HLogInputFormat.class); diff --git src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java index 9b1f239..b0a7b69 100644 --- src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java +++ src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java @@ -23,8 +23,6 @@ import java.text.SimpleDateFormat; import java.util.Map; import java.util.TreeMap; -import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.fs.Path; @@ -57,8 +55,6 @@ import org.apache.hadoop.util.ToolRunner; * WAL replay can also generate HFiles for later bulk importing, * in that case the WAL is replayed for a single table only. */ -@InterfaceAudience.Public -@InterfaceStability.Stable public class WALPlayer extends Configured implements Tool { final static String NAME = "WALPlayer"; final static String BULK_OUTPUT_CONF_KEY = "hlog.bulk.output"; diff --git src/main/java/org/apache/hadoop/hbase/regionserver/metrics/OperationMetrics.java src/main/java/org/apache/hadoop/hbase/regionserver/metrics/OperationMetrics.java index 305dfab..fa91293 100644 --- src/main/java/org/apache/hadoop/hbase/regionserver/metrics/OperationMetrics.java +++ src/main/java/org/apache/hadoop/hbase/regionserver/metrics/OperationMetrics.java @@ -20,7 +20,6 @@ package org.apache.hadoop.hbase.regionserver.metrics; import java.util.Set; -import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.client.Append; @@ -37,7 +36,6 @@ import org.apache.hadoop.hbase.util.Bytes; * metrics are stored in {@link RegionMetricsStorage} and exposed to hadoop * metrics through {@link RegionServerDynamicMetrics}. */ -@InterfaceAudience.Private public class OperationMetrics { private static final String DELETE_KEY = "delete_"; diff --git src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionMetricsStorage.java src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionMetricsStorage.java index 5d4beff..3239540 100644 --- src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionMetricsStorage.java +++ src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionMetricsStorage.java @@ -25,7 +25,6 @@ import java.util.concurrent.ConcurrentMap; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicLong; -import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Pair; /** @@ -34,7 +33,6 @@ import org.apache.hadoop.hbase.util.Pair; * numbers. These numbers are exposed to Hadoop metrics through * RegionServerDynamicMetrics. */ -@InterfaceAudience.Private public class RegionMetricsStorage { // for simple numeric metrics (# of blocks read from block cache) diff --git src/main/java/org/apache/hadoop/hbase/regionserver/wal/CompressionContext.java src/main/java/org/apache/hadoop/hbase/regionserver/wal/CompressionContext.java index 5a877af..10aec00 100644 --- src/main/java/org/apache/hadoop/hbase/regionserver/wal/CompressionContext.java +++ src/main/java/org/apache/hadoop/hbase/regionserver/wal/CompressionContext.java @@ -20,12 +20,10 @@ package org.apache.hadoop.hbase.regionserver.wal; import java.lang.reflect.Constructor; import java.lang.reflect.InvocationTargetException; -import org.apache.hadoop.classification.InterfaceAudience; /** * Context that holds the various dictionaries for compression in HLog. */ -@InterfaceAudience.Private class CompressionContext { final Dictionary regionDict; final Dictionary tableDict; diff --git src/main/java/org/apache/hadoop/hbase/regionserver/wal/Compressor.java src/main/java/org/apache/hadoop/hbase/regionserver/wal/Compressor.java index 8a0dbaf..8193e1f 100644 --- src/main/java/org/apache/hadoop/hbase/regionserver/wal/Compressor.java +++ src/main/java/org/apache/hadoop/hbase/regionserver/wal/Compressor.java @@ -17,7 +17,6 @@ */ package org.apache.hadoop.hbase.regionserver.wal; -import org.apache.hadoop.classification.InterfaceAudience; import java.io.DataInput; import java.io.DataOutput; @@ -37,7 +36,6 @@ import com.google.common.base.Preconditions; * A set of static functions for running our custom WAL compression/decompression. * Also contains a command line tool to compress and uncompress HLogs. */ -@InterfaceAudience.Private public class Compressor { /** * Command line tool to compress and uncompress WALs. diff --git src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java index e1cfed1..5dbf3bf 100644 --- src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java +++ src/main/java/org/apache/hadoop/hbase/regionserver/wal/Dictionary.java @@ -18,7 +18,6 @@ package org.apache.hadoop.hbase.regionserver.wal; -import org.apache.hadoop.classification.InterfaceAudience; /** * Dictionary interface @@ -26,7 +25,6 @@ import org.apache.hadoop.classification.InterfaceAudience; * Dictionary indexes should be either bytes or shorts, only positive. (The * first bit is reserved for detecting whether something is compressed or not). */ -@InterfaceAudience.Private interface Dictionary { static final byte NOT_IN_DICTIONARY = -1; diff --git src/main/java/org/apache/hadoop/hbase/regionserver/wal/LRUDictionary.java src/main/java/org/apache/hadoop/hbase/regionserver/wal/LRUDictionary.java index 6e0b20b..5ee18fc 100644 --- src/main/java/org/apache/hadoop/hbase/regionserver/wal/LRUDictionary.java +++ src/main/java/org/apache/hadoop/hbase/regionserver/wal/LRUDictionary.java @@ -20,7 +20,6 @@ package org.apache.hadoop.hbase.regionserver.wal; import java.util.HashMap; -import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; @@ -33,7 +32,6 @@ import com.google.common.base.Preconditions; * (2 ^ 15) * 5 (Regionname, Row key, CF, Column qual, table) * 100 bytes (these are some big names) = ~16MB. * If you want to get silly, even at 1kb entries, it maxes out at 160 megabytes. */ -@InterfaceAudience.Private public class LRUDictionary implements Dictionary { private final BidirectionalLRUMap backingStore = new BidirectionalLRUMap(); diff --git src/test/java/org/apache/hadoop/hbase/regionserver/wal/HLogPerformanceEvaluation.java src/test/java/org/apache/hadoop/hbase/regionserver/wal/HLogPerformanceEvaluation.java index 0258cda..e534cc2 100644 --- src/test/java/org/apache/hadoop/hbase/regionserver/wal/HLogPerformanceEvaluation.java +++ src/test/java/org/apache/hadoop/hbase/regionserver/wal/HLogPerformanceEvaluation.java @@ -32,7 +32,6 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; import org.apache.hadoop.conf.Configured; -import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HColumnDescriptor; @@ -44,16 +43,13 @@ import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.regionserver.HRegion; -import org.apache.hadoop.hbase.regionserver.wal.HLog; import org.apache.hadoop.hbase.regionserver.wal.HLog.Entry; -import org.apache.hadoop.hbase.regionserver.wal.WALEdit; /** * This class runs performance benchmarks for {@link HLog}. * See usage for this tool by running: * $ hbase org.apache.hadoop.hbase.regionserver.wal.HLogPerformanceEvaluation -h */ -@InterfaceAudience.Private public final class HLogPerformanceEvaluation extends Configured implements Tool { static final Log LOG = LogFactory.getLog(HLogPerformanceEvaluation.class.getName());