diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java index 3d40c70..118ee08 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java @@ -35,7 +35,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallScanner.java index f9bdd55..4a0df33 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallScanner.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallScanner.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.client; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java index e43a712..89e9ec5 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java @@ -21,9 +21,9 @@ package org.apache.hadoop.hbase.client; import static org.apache.hadoop.hbase.client.MetricsConnection.CLIENT_SIDE_METRICS_ENABLED_KEY; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import java.io.Closeable; import java.io.IOException; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/FlushRegionCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/FlushRegionCallable.java index 73bdb74..cee2779 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/FlushRegionCallable.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/FlushRegionCallable.java @@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.FlushRegionRespons import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * A Callable for flushRegion() RPC. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java index 9541967..510e92a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java @@ -177,8 +177,8 @@ import org.apache.hadoop.util.StringUtils; import org.apache.zookeeper.KeeperException; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ByteString; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * HBaseAdmin is no longer a client API. It is marked InterfaceAudience.Private indicating that diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java index befc671..cf62561 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java @@ -69,10 +69,10 @@ import org.apache.hadoop.hbase.util.ReflectionUtils; import org.apache.hadoop.hbase.util.Threads; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * An implementation of {@link Table}. Used to communicate with a single HBase table. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MetricsConnection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MetricsConnection.java index 53a3326..303611d 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MetricsConnection.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MetricsConnection.java @@ -18,8 +18,8 @@ package org.apache.hadoop.hbase.client; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; import com.codahale.metrics.Counter; import com.codahale.metrics.Histogram; import com.codahale.metrics.MetricRegistry; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiServerCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiServerCallable.java index f78f348..e78a7c9 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiServerCallable.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/MultiServerCallable.java @@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionAction; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Callable that handles the multi method call going against a single diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionCoprocessorServiceExec.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionCoprocessorServiceExec.java index ad1d2a1..1101020 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionCoprocessorServiceExec.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionCoprocessorServiceExec.java @@ -23,8 +23,8 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Objects; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java index 6ce4956..6d57aa7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.ExceptionUtil; import org.apache.hadoop.ipc.RemoteException; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Runs an rpc'ing {@link RetryingCallable}. Sets into rpc client diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java index f4e2614..7745710 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java @@ -48,7 +48,7 @@ import org.apache.hadoop.hbase.protobuf.RequestConverter; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java index 72d69ec..a81306e 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java @@ -52,8 +52,8 @@ import org.apache.hadoop.hbase.regionserver.RegionServerStoppedException; import org.apache.hadoop.ipc.RemoteException; import org.apache.hadoop.net.DNS; -import com.google.protobuf.ServiceException; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; /** * Scanner operations such as create, next, etc. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java index 3e9db00..4b0dada 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java @@ -32,10 +32,10 @@ import org.apache.hadoop.hbase.client.coprocessor.Batch; import org.apache.hadoop.hbase.filter.CompareFilter; import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Used to communicate with a single HBase table. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java index 5d4ac8e..deb700b 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/TableState.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.client; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java index 594a459..feb4c8b 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java @@ -53,8 +53,8 @@ import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateServi import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Pair; -import com.google.protobuf.ByteString; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * This client class is for invoking the aggregate functions deployed on the diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationSerDeHelper.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationSerDeHelper.java index 9682f89..fb1dcff 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationSerDeHelper.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationSerDeHelper.java @@ -18,7 +18,7 @@ */ package org.apache.hadoop.hbase.client.replication; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.TableName; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java index e247c08..283244d 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java @@ -24,7 +24,7 @@ import java.io.IOException; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.classification.InterfaceAudience; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * Defines how value for specific column is interpreted and provides utility diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java index 3cbb7b9..1bc8af3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos; import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A binary comparator which lexicographically compares against the specified diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java index a26edbc..d001532 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos; import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A comparator which compares against a specified byte array, but only compares diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java index db51df7..830417e 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A bit comparator which performs the specified bitwise operation on each of the bytes diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java index fd65130..1c9429f 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Simple filter that returns first N columns on row only. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java index e5ec412..c9da5b2 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A filter, based on the ColumnCountGetFilter, takes two arguments: limit and offset. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java index ff6e8e2..ea84042 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java @@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used for selecting only those keys with columns that matches diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java index 04682c5..75324d4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used for selecting only those keys with columns that are diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java index f7c6f26..5711d3c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A filter for adding inter-column timestamp matching diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java index b3f9a1a..aabf6d1 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** *

diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java index da7a084..2efde28 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Implementation of {@link Filter} that represents an ordered List of Filters diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java index 4d7a18a..5a2d2d4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This is a Filter wrapper class which is used in the server side. Some filter diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java index 80a1deb..4f3e605 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A filter that will only return the first KV from each row. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java index 2e9510f..5b93c97 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java @@ -30,8 +30,8 @@ import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.ByteString; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * The filter looks for the given columns in KeyValue. Once there is a match for diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java index 500d01d..cfc9d5c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java @@ -38,7 +38,7 @@ import org.apache.hadoop.hbase.util.UnsafeAccess; import org.apache.hadoop.hbase.util.UnsafeAvailChecker; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This is optimized version of a standard FuzzyRowFilter Filters data based on fuzzy row key. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java index 1096f5e..acc05e9 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A Filter that stops after the given row. There is no "RowStopFilter" because diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java index 2fd5aba..0da93a2 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A filter that will only return the key component of each KV (the value will diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java index 9c56772..48027f4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.filter; import java.nio.ByteBuffer; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java index 5f9c833..31456c7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java @@ -33,7 +33,7 @@ import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Filter to support scan multiple row key ranges. It can construct the row key ranges from the diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java index 0c14649..e7d08bb 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used for selecting only those keys with columns that matches diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java index 160232f..5d7ad83 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A binary comparator which lexicographically compares against the specified diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java index adc9c54..6e356cd 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Implementation of Filter interface that limits results to a specific page * size. It terminates scanning once the number of filter-passed rows is > diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java index d09ea2c..43d400f 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Pass results that have same row prefix. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java index 3aa3558..b148ab3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used to filter based on the column qualifier. It takes an diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java index decdc78..9af8fd0 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A filter that includes rows based on a chance. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java index 70dd1f9..87881ee 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java @@ -38,7 +38,7 @@ import org.joni.Option; import org.joni.Regex; import org.joni.Syntax; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This comparator is for use with {@link CompareFilter} implementations, such diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java index 559eff8..3e52007 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used to filter based on the key. It takes an operator diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java index d030fd2..8b138f2 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java @@ -33,7 +33,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A {@link Filter} that checks a single column value, but does not emit the diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java index df4e482..aac6754 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java @@ -37,7 +37,7 @@ import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used to filter cells based on value. It takes a {@link CompareFilter.CompareOp} diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java index 3aced13..1d39dc3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A wrapper filter that filters an entire row if any of the Cell checks do diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java index 63fd0a3..c07114a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java index f0e5afe..e48f7c5 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java @@ -30,7 +30,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import com.google.common.base.Preconditions; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Filter that returns only cells whose timestamp (version) is diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java index 2f679f0..e27e7ce 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This filter is used to filter based on column value. It takes an diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java index e75ca49..3fb119c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A wrapper filter that returns true from {@link #filterAllRemaining()} as soon diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java index ec6332a..7d702f6 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java @@ -19,11 +19,11 @@ package org.apache.hadoop.hbase.ipc; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import java.io.IOException; import java.net.ConnectException; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncCall.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncCall.java index a5da0dc..71bb74f 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncCall.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncCall.java @@ -17,8 +17,8 @@ */ package org.apache.hadoop.hbase.ipc; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; import io.netty.channel.EventLoop; import io.netty.util.concurrent.DefaultPromise; import org.apache.commons.logging.Log; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcChannel.java index 53eb824..ecb8887 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcChannel.java @@ -60,9 +60,9 @@ import org.apache.hadoop.security.token.TokenSelector; import org.apache.htrace.Span; import org.apache.htrace.Trace; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; import io.netty.bootstrap.Bootstrap; import io.netty.buffer.ByteBuf; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java index c2bd457..f9b8cb2 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java @@ -61,11 +61,11 @@ import org.apache.hadoop.hbase.util.PoolMap; import org.apache.hadoop.hbase.util.Threads; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcChannel; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Netty client for the requests and responses diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncServerResponseHandler.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncServerResponseHandler.java index e0c7586..a2bdd9f 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncServerResponseHandler.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncServerResponseHandler.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.RPCProtos; import org.apache.hadoop.ipc.RemoteException; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; import io.netty.buffer.ByteBuf; import io.netty.buffer.ByteBufInputStream; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/BlockingRpcCallback.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/BlockingRpcCallback.java index 3aa59c7..c7e507b 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/BlockingRpcCallback.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/BlockingRpcCallback.java @@ -23,7 +23,7 @@ import java.io.InterruptedIOException; import org.apache.hadoop.hbase.classification.InterfaceAudience; -import com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; /** * Simple {@link RpcCallback} implementation providing a diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/Call.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/Call.java index 5f90837..4e88ad0 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/Call.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/Call.java @@ -17,8 +17,8 @@ */ package org.apache.hadoop.hbase.ipc; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; import org.apache.hadoop.hbase.CellScanner; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.MetricsConnection; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.java index b1d54a4..bb751e4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.java @@ -26,13 +26,13 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.protobuf.ResponseConverter; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcChannel; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Base class which provides clients with an RPC connection to diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java index d98d81d..f48f5ec 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java @@ -46,8 +46,8 @@ import org.apache.hadoop.io.compress.Compressor; import org.apache.hadoop.io.compress.Decompressor; import com.google.common.base.Preconditions; -import com.google.protobuf.CodedOutputStream; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * Utility to help ipc'ing. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java index 6e59972..7f729b7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java @@ -30,9 +30,9 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceResponse; import org.apache.hadoop.hbase.util.ByteStringer; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Provides clients with an RPC connection to call coprocessor endpoint {@link com.google.protobuf.Service}s diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java index 321dd62..f722c87 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java @@ -33,9 +33,9 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServic import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Provides clients with an RPC connection to call coprocessor endpoint {@link com.google.protobuf.Service}s diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionServerCoprocessorRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionServerCoprocessorRpcChannel.java index 24d2de4..6baccef 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionServerCoprocessorRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionServerCoprocessorRpcChannel.java @@ -24,9 +24,9 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceResponse; import org.apache.hadoop.hbase.util.ByteStringer; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Provides clients with an RPC connection to call coprocessor endpoint diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java index 540e224..d3865c5 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.ipc; -import com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.security.User; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClientImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClientImpl.java index 83d4adf..9a266d8 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClientImpl.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClientImpl.java @@ -97,10 +97,10 @@ import org.apache.htrace.Trace; import org.apache.htrace.TraceScope; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.Message.Builder; -import com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message.Builder; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; /** * Does RPC against a cluster. Manages connections per regionserver in the cluster. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java index aa407f7..c058533 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java @@ -23,8 +23,8 @@ import java.io.IOException; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.util.StringUtils; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Used for server-side protobuf RPC service invocations. This handler allows diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/TimeLimitedRpcController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/TimeLimitedRpcController.java index de502cb..dd6431e 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/TimeLimitedRpcController.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/TimeLimitedRpcController.java @@ -23,8 +23,8 @@ import java.util.concurrent.atomic.AtomicReference; import org.apache.hadoop.hbase.classification.InterfaceAudience; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; @InterfaceAudience.Private public class TimeLimitedRpcController implements RpcController { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 90516ec..9b4ba84 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -167,16 +167,16 @@ import com.google.common.collect.ArrayListMultimap; import com.google.common.collect.ListMultimap; import com.google.common.collect.Lists; import com.google.common.net.HostAndPort; -import com.google.protobuf.ByteString; -import com.google.protobuf.CodedInputStream; -import com.google.protobuf.InvalidProtocolBufferException; -import com.google.protobuf.Message; -import com.google.protobuf.Parser; -import com.google.protobuf.RpcChannel; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Parser; +import org.apache.hbase.shaded.com.google.protobuf.RpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; /** * Protobufs utility. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java index 45d15a3..d45eddd 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java @@ -115,7 +115,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.Pair; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Helper utility to build protocol buffer requests, diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ResponseConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ResponseConverter.java index 421907d..5160b44 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ResponseConverter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ResponseConverter.java @@ -59,8 +59,8 @@ import org.apache.hadoop.hbase.regionserver.RegionOpeningState; import org.apache.hadoop.hbase.security.access.UserPermission; import org.apache.hadoop.util.StringUtils; -import com.google.protobuf.ByteString; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Helper utility to build protocol buffer responses, diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenIdentifier.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenIdentifier.java index 64e1a39..9a03a0c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenIdentifier.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenIdentifier.java @@ -29,7 +29,7 @@ import org.apache.hadoop.io.Text; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.token.TokenIdentifier; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Represents the identity information stored in an HBase authentication token. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java index 7527049..bb3107c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java @@ -46,7 +46,7 @@ import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.Visibil import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Utility client for doing visibility labels admin operations. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterAddressTracker.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterAddressTracker.java index 6f4859a..dc8d438 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterAddressTracker.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterAddressTracker.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos; import org.apache.zookeeper.KeeperException; import org.apache.zookeeper.data.Stat; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Manages the location of the current active Master for the RegionServer. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java index 0b53f95..3fcc345 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java @@ -58,7 +58,7 @@ import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.ipc.RemoteException; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Utility class to perform operation (get/wait for/verify/set/delete) on znode in ZooKeeper diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java index 9e01d09..53370b0 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java @@ -77,7 +77,7 @@ import org.apache.zookeeper.proto.DeleteRequest; import org.apache.zookeeper.proto.SetDataRequest; import org.apache.zookeeper.server.ZooKeeperSaslServer; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Internal HBase utility class for ZooKeeper. diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java index 0a5a37f..29c601c 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java @@ -86,9 +86,9 @@ import org.junit.experimental.categories.Category; import org.mockito.Mockito; import com.google.common.base.Stopwatch; -import com.google.protobuf.ByteString; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test client behavior w/o setting up a cluster. diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestMetricsConnection.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestMetricsConnection.java index 5191880..c45de59 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestMetricsConnection.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestMetricsConnection.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.client; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetRequest; diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java index 4d55c33..f3f5028 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java @@ -41,7 +41,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.mockito.Mockito; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** * Test snapshot logic from the client diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/exceptions/TestClientExceptionsUtil.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/exceptions/TestClientExceptionsUtil.java index 968e55c..e44611c 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/exceptions/TestClientExceptionsUtil.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/exceptions/TestClientExceptionsUtil.java @@ -18,7 +18,7 @@ */ package org.apache.hadoop.hbase.exceptions; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.junit.Test; import java.io.IOException; diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java index c0c43ed..7e0cd5a 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase; import com.google.common.net.HostAndPort; import com.google.common.net.InetAddresses; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import java.io.Serializable; import java.util.ArrayList; diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/types/PBType.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/types/PBType.java index 3d545f6..1d732b9 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/types/PBType.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/types/PBType.java @@ -22,9 +22,9 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Order; import org.apache.hadoop.hbase.util.PositionedByteRange; -import com.google.protobuf.CodedInputStream; -import com.google.protobuf.CodedOutputStream; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * A base-class for {@link DataType} implementations backed by protobuf. See diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java index 7b9eb0b..593ea17 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java @@ -35,7 +35,7 @@ import java.util.Comparator; import java.util.Iterator; import java.util.List; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java index c9ab23c..b111be1 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java @@ -51,9 +51,9 @@ import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.RegionScanner; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Defines a protocol to delete data in bulk based on a scan. The scan can be range scan or with diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java index 4309cdc..188d8d0 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java @@ -36,9 +36,9 @@ import org.apache.hadoop.hbase.protobuf.ResponseConverter; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Sample coprocessor endpoint exposing a Service interface for counting rows and key values. diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java index 373e036..51f66bb 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.coprocessor.example.generated; public final class BulkDeleteProtos { private BulkDeleteProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface BulkDeleteRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required .hbase.pb.Scan scan = 1; /** @@ -59,14 +59,14 @@ public final class BulkDeleteProtos { * Protobuf type {@code hbase.pb.BulkDeleteRequest} */ public static final class BulkDeleteRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements BulkDeleteRequestOrBuilder { // Use BulkDeleteRequest.newBuilder() to construct. - private BulkDeleteRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private BulkDeleteRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private BulkDeleteRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private BulkDeleteRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final BulkDeleteRequest defaultInstance; public static BulkDeleteRequest getDefaultInstance() { @@ -77,20 +77,20 @@ public final class BulkDeleteProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BulkDeleteRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -142,40 +142,40 @@ public final class BulkDeleteProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public BulkDeleteRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new BulkDeleteRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -183,7 +183,7 @@ public final class BulkDeleteProtos { * Protobuf enum {@code hbase.pb.BulkDeleteRequest.DeleteType} */ public enum DeleteType - implements com.google.protobuf.ProtocolMessageEnum { + implements org.apache.hbase.shaded.com.google.protobuf.ProtocolMessageEnum { /** * ROW = 0; */ @@ -232,27 +232,27 @@ public final class BulkDeleteProtos { } } - public static com.google.protobuf.Internal.EnumLiteMap + public static org.apache.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } - private static com.google.protobuf.Internal.EnumLiteMap + private static org.apache.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap() { + new org.apache.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap() { public DeleteType findValueByNumber(int number) { return DeleteType.valueOf(number); } }; - public final com.google.protobuf.Descriptors.EnumValueDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } - public final com.google.protobuf.Descriptors.EnumDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } - public static final com.google.protobuf.Descriptors.EnumDescriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.getDescriptor().getEnumTypes().get(0); } @@ -260,7 +260,7 @@ public final class BulkDeleteProtos { private static final DeleteType[] VALUES = values(); public static DeleteType valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); @@ -381,7 +381,7 @@ public final class BulkDeleteProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -406,19 +406,19 @@ public final class BulkDeleteProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(1, scan_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeEnumSize(2, deleteType_.getNumber()); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, timestamp_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeUInt32Size(4, rowBatchSize_); } size += getUnknownFields().getSerializedSize(); @@ -499,24 +499,24 @@ public final class BulkDeleteProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(java.io.InputStream input) @@ -525,7 +525,7 @@ public final class BulkDeleteProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -535,18 +535,18 @@ public final class BulkDeleteProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -560,7 +560,7 @@ public final class BulkDeleteProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -568,14 +568,14 @@ public final class BulkDeleteProtos { * Protobuf type {@code hbase.pb.BulkDeleteRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -588,12 +588,12 @@ public final class BulkDeleteProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getScanFieldBuilder(); } } @@ -622,7 +622,7 @@ public final class BulkDeleteProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_descriptor; } @@ -668,7 +668,7 @@ public final class BulkDeleteProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest)other); } else { @@ -716,13 +716,13 @@ public final class BulkDeleteProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -736,7 +736,7 @@ public final class BulkDeleteProtos { // required .hbase.pb.Scan scan = 1; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> scanBuilder_; /** * required .hbase.pb.Scan scan = 1; @@ -837,11 +837,11 @@ public final class BulkDeleteProtos { /** * required .hbase.pb.Scan scan = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> getScanFieldBuilder() { if (scanBuilder_ == null) { - scanBuilder_ = new com.google.protobuf.SingleFieldBuilder< + scanBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder>( scan_, getParentForChildren(), @@ -965,7 +965,7 @@ public final class BulkDeleteProtos { } public interface BulkDeleteResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required uint64 rowsDeleted = 1; /** @@ -991,14 +991,14 @@ public final class BulkDeleteProtos { * Protobuf type {@code hbase.pb.BulkDeleteResponse} */ public static final class BulkDeleteResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements BulkDeleteResponseOrBuilder { // Use BulkDeleteResponse.newBuilder() to construct. - private BulkDeleteResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private BulkDeleteResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private BulkDeleteResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private BulkDeleteResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final BulkDeleteResponse defaultInstance; public static BulkDeleteResponse getDefaultInstance() { @@ -1009,20 +1009,20 @@ public final class BulkDeleteProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BulkDeleteResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1050,40 +1050,40 @@ public final class BulkDeleteProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public BulkDeleteResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new BulkDeleteResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -1137,7 +1137,7 @@ public final class BulkDeleteProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1156,11 +1156,11 @@ public final class BulkDeleteProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, rowsDeleted_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, versionsDeleted_); } size += getUnknownFields().getSerializedSize(); @@ -1223,24 +1223,24 @@ public final class BulkDeleteProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(java.io.InputStream input) @@ -1249,7 +1249,7 @@ public final class BulkDeleteProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1259,18 +1259,18 @@ public final class BulkDeleteProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1284,7 +1284,7 @@ public final class BulkDeleteProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1292,14 +1292,14 @@ public final class BulkDeleteProtos { * Protobuf type {@code hbase.pb.BulkDeleteResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1312,12 +1312,12 @@ public final class BulkDeleteProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1337,7 +1337,7 @@ public final class BulkDeleteProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_descriptor; } @@ -1371,7 +1371,7 @@ public final class BulkDeleteProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse)other); } else { @@ -1401,13 +1401,13 @@ public final class BulkDeleteProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -1500,7 +1500,7 @@ public final class BulkDeleteProtos { * Protobuf service {@code hbase.pb.BulkDeleteService} */ public static abstract class BulkDeleteService - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected BulkDeleteService() {} public interface Interface { @@ -1508,39 +1508,39 @@ public final class BulkDeleteProtos { * rpc delete(.hbase.pb.BulkDeleteRequest) returns (.hbase.pb.BulkDeleteResponse); */ public abstract void delete( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new BulkDeleteService() { @java.lang.Override public void delete( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.delete(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -1554,9 +1554,9 @@ public final class BulkDeleteProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1570,9 +1570,9 @@ public final class BulkDeleteProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1593,26 +1593,26 @@ public final class BulkDeleteProtos { * rpc delete(.hbase.pb.BulkDeleteRequest) returns (.hbase.pb.BulkDeleteResponse); */ public abstract void delete( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -1621,7 +1621,7 @@ public final class BulkDeleteProtos { switch(method.getIndex()) { case 0: this.delete(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -1629,9 +1629,9 @@ public final class BulkDeleteProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1645,9 +1645,9 @@ public final class BulkDeleteProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1662,31 +1662,31 @@ public final class BulkDeleteProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteService implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void delete( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance())); @@ -1694,28 +1694,28 @@ public final class BulkDeleteProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse delete( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse delete( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -1728,22 +1728,22 @@ public final class BulkDeleteProtos { // @@protoc_insertion_point(class_scope:hbase.pb.BulkDeleteService) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BulkDeleteRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BulkDeleteResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1761,29 +1761,29 @@ public final class BulkDeleteProtos { ".coprocessor.example.generatedB\020BulkDele" + "teProtosH\001\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_hbase_pb_BulkDeleteRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_BulkDeleteRequest_descriptor, new java.lang.String[] { "Scan", "DeleteType", "Timestamp", "RowBatchSize", }); internal_static_hbase_pb_BulkDeleteResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_BulkDeleteResponse_descriptor, new java.lang.String[] { "RowsDeleted", "VersionsDeleted", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(), }, assigner); } diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java index b780985..6ca5164 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java @@ -6,23 +6,23 @@ package org.apache.hadoop.hbase.coprocessor.example.generated; public final class ExampleProtos { private ExampleProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface CountRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.CountRequest} */ public static final class CountRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements CountRequestOrBuilder { // Use CountRequest.newBuilder() to construct. - private CountRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private CountRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private CountRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private CountRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final CountRequest defaultInstance; public static CountRequest getDefaultInstance() { @@ -33,19 +33,19 @@ public final class ExampleProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CountRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -63,40 +63,40 @@ public final class ExampleProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public CountRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new CountRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -111,7 +111,7 @@ public final class ExampleProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -165,24 +165,24 @@ public final class ExampleProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(java.io.InputStream input) @@ -191,7 +191,7 @@ public final class ExampleProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -201,18 +201,18 @@ public final class ExampleProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -226,7 +226,7 @@ public final class ExampleProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -234,14 +234,14 @@ public final class ExampleProtos { * Protobuf type {@code hbase.pb.CountRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -254,12 +254,12 @@ public final class ExampleProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -275,7 +275,7 @@ public final class ExampleProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_descriptor; } @@ -298,7 +298,7 @@ public final class ExampleProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)other); } else { @@ -318,13 +318,13 @@ public final class ExampleProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -347,7 +347,7 @@ public final class ExampleProtos { } public interface CountResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int64 count = 1 [default = 0]; /** @@ -363,14 +363,14 @@ public final class ExampleProtos { * Protobuf type {@code hbase.pb.CountResponse} */ public static final class CountResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements CountResponseOrBuilder { // Use CountResponse.newBuilder() to construct. - private CountResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private CountResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private CountResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private CountResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final CountResponse defaultInstance; public static CountResponse getDefaultInstance() { @@ -381,20 +381,20 @@ public final class ExampleProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CountResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -417,40 +417,40 @@ public final class ExampleProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public CountResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new CountResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -487,7 +487,7 @@ public final class ExampleProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -503,7 +503,7 @@ public final class ExampleProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(1, count_); } size += getUnknownFields().getSerializedSize(); @@ -557,24 +557,24 @@ public final class ExampleProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(java.io.InputStream input) @@ -583,7 +583,7 @@ public final class ExampleProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -593,18 +593,18 @@ public final class ExampleProtos { } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -618,7 +618,7 @@ public final class ExampleProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -626,14 +626,14 @@ public final class ExampleProtos { * Protobuf type {@code hbase.pb.CountResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -646,12 +646,12 @@ public final class ExampleProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -669,7 +669,7 @@ public final class ExampleProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_descriptor; } @@ -699,7 +699,7 @@ public final class ExampleProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse)other); } else { @@ -726,13 +726,13 @@ public final class ExampleProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -792,7 +792,7 @@ public final class ExampleProtos { * Protobuf service {@code hbase.pb.RowCountService} */ public static abstract class RowCountService - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected RowCountService() {} public interface Interface { @@ -800,55 +800,55 @@ public final class ExampleProtos { * rpc getRowCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse); */ public abstract void getRowCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc getKeyValueCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse); */ public abstract void getKeyValueCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new RowCountService() { @java.lang.Override public void getRowCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.getRowCount(controller, request, done); } @java.lang.Override public void getKeyValueCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.getKeyValueCount(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -864,9 +864,9 @@ public final class ExampleProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -882,9 +882,9 @@ public final class ExampleProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -907,34 +907,34 @@ public final class ExampleProtos { * rpc getRowCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse); */ public abstract void getRowCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc getKeyValueCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse); */ public abstract void getKeyValueCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -943,12 +943,12 @@ public final class ExampleProtos { switch(method.getIndex()) { case 0: this.getRowCount(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.getKeyValueCount(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -956,9 +956,9 @@ public final class ExampleProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -974,9 +974,9 @@ public final class ExampleProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -993,46 +993,46 @@ public final class ExampleProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.RowCountService implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void getRowCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance())); } public void getKeyValueCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance())); @@ -1040,33 +1040,33 @@ public final class ExampleProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getRowCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getKeyValueCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getRowCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -1076,9 +1076,9 @@ public final class ExampleProtos { public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getKeyValueCount( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, @@ -1091,22 +1091,22 @@ public final class ExampleProtos { // @@protoc_insertion_point(class_scope:hbase.pb.RowCountService) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_CountRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_hbase_pb_CountRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_CountResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_hbase_pb_CountResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1119,29 +1119,29 @@ public final class ExampleProtos { "ache.hadoop.hbase.coprocessor.example.ge" + "neratedB\rExampleProtosH\001\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_hbase_pb_CountRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_CountRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_CountRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_CountResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_CountResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_CountResponse_descriptor, new java.lang.String[] { "Count", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/types/PBCell.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/types/PBCell.java index c063aa9..29f2e4d 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/types/PBCell.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/types/PBCell.java @@ -17,8 +17,8 @@ */ package org.apache.hadoop.hbase.types; -import com.google.protobuf.CodedInputStream; -import com.google.protobuf.CodedOutputStream; +import org.apache.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream; import org.apache.hadoop.hbase.protobuf.generated.CellProtos; import org.apache.hadoop.hbase.util.PositionedByteRange; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/ipc/IntegrationTestRpcClient.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/ipc/IntegrationTestRpcClient.java index c28f3e6..f1a85c7 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/ipc/IntegrationTestRpcClient.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/ipc/IntegrationTestRpcClient.java @@ -61,11 +61,11 @@ import org.junit.Ignore; import org.junit.Test; import org.junit.experimental.categories.Category; import com.google.common.collect.Lists; -import com.google.protobuf.BlockingService; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; -import com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; @Category(IntegrationTests.class) public class IntegrationTestRpcClient { diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java index 2cc67be..025ac31 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/Procedure.java @@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.util.NonceKey; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Base Procedure class responsible to handle the Procedure Metadata diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureException.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureException.java index 71aae84..c3ffe62 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureException.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureException.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage; import org.apache.hadoop.hbase.util.ForeignExceptionUtil; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A RemoteProcedureException is an exception from another thread or process. diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormat.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormat.java index add7d03..e1eab1a 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormat.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormat.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos.ProcedureWALEn import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos.ProcedureWALHeader; import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos.ProcedureWALTrailer; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Helper class that contains the WAL serialization utils. diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormatReader.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormatReader.java index 312eedb..b8285bb 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormatReader.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALFormatReader.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.procedure2.store.ProcedureStoreTracker; import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos; import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos.ProcedureWALEntry; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Helper class that loads the procedures stored in a WAL diff --git a/hbase-protocol/pom.xml b/hbase-protocol/pom.xml index d43b7b9..426bc5e 100644 --- a/hbase-protocol/pom.xml +++ b/hbase-protocol/pom.xml @@ -37,6 +37,54 @@ + + + org.apache.maven.plugins + maven-shade-plugin + 2.4.3 + + + package + + shade + + + + + com.google.protobuf + org.apache.hbase.shaded.com.google.protobuf + + + + + + commons-logging:commons-logging + com.github.stephenc.findbugs:findbugs-annotations + log4j:log4j + org.hamcrest:hamcrest-core + org.mockito:mockito-all + junit:junit + org.apache.hbase:hbase-annotations + + + + + + + + org.apache.maven.plugins maven-site-plugin @@ -106,6 +154,9 @@ + org.apache.hbase @@ -121,6 +172,7 @@ com.google.protobuf protobuf-java + 2.5.0 commons-logging diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java index 2a30e99..719cb84 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java @@ -67,10 +67,10 @@ import org.apache.hadoop.hbase.rest.model.TableSchemaModel; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.util.StringUtils; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * HTable interface to remote tables accessed via REST gateway diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java index 7512d3e..05531a2 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java @@ -77,7 +77,7 @@ import org.apache.hadoop.hbase.util.Base64; import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; import com.sun.jersey.api.json.JSONConfiguration; import com.sun.jersey.api.json.JSONJAXBContext; import com.sun.jersey.api.json.JSONMarshaller; diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellMessage.java index 4c859e1..b13a1ff 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class CellMessage { private CellMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface CellOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional bytes row = 1; /** @@ -27,7 +27,7 @@ public final class CellMessage { * unused if Cell is in a CellSet * */ - com.google.protobuf.ByteString getRow(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getRow(); // optional bytes column = 2; /** @@ -37,7 +37,7 @@ public final class CellMessage { /** * optional bytes column = 2; */ - com.google.protobuf.ByteString getColumn(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getColumn(); // optional int64 timestamp = 3; /** @@ -57,20 +57,20 @@ public final class CellMessage { /** * optional bytes data = 4; */ - com.google.protobuf.ByteString getData(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getData(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Cell} */ public static final class Cell extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements CellOrBuilder { // Use Cell.newBuilder() to construct. - private Cell(com.google.protobuf.GeneratedMessage.Builder builder) { + private Cell(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Cell(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Cell(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Cell defaultInstance; public static Cell getDefaultInstance() { @@ -81,20 +81,20 @@ public final class CellMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Cell( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -132,47 +132,47 @@ public final class CellMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell.class, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Cell parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Cell(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // optional bytes row = 1; public static final int ROW_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString row_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_; /** * optional bytes row = 1; * @@ -190,13 +190,13 @@ public final class CellMessage { * unused if Cell is in a CellSet * */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } // optional bytes column = 2; public static final int COLUMN_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString column_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString column_; /** * optional bytes column = 2; */ @@ -206,7 +206,7 @@ public final class CellMessage { /** * optional bytes column = 2; */ - public com.google.protobuf.ByteString getColumn() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getColumn() { return column_; } @@ -228,7 +228,7 @@ public final class CellMessage { // optional bytes data = 4; public static final int DATA_FIELD_NUMBER = 4; - private com.google.protobuf.ByteString data_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString data_; /** * optional bytes data = 4; */ @@ -238,15 +238,15 @@ public final class CellMessage { /** * optional bytes data = 4; */ - public com.google.protobuf.ByteString getData() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getData() { return data_; } private void initFields() { - row_ = com.google.protobuf.ByteString.EMPTY; - column_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + column_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; timestamp_ = 0L; - data_ = com.google.protobuf.ByteString.EMPTY; + data_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -257,7 +257,7 @@ public final class CellMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -282,19 +282,19 @@ public final class CellMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, row_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, column_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(3, timestamp_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(4, data_); } size += getUnknownFields().getSerializedSize(); @@ -310,24 +310,24 @@ public final class CellMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom(java.io.InputStream input) @@ -336,7 +336,7 @@ public final class CellMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -346,18 +346,18 @@ public final class CellMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -371,7 +371,7 @@ public final class CellMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -379,14 +379,14 @@ public final class CellMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Cell} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.CellOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -399,12 +399,12 @@ public final class CellMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -413,13 +413,13 @@ public final class CellMessage { public Builder clear() { super.clear(); - row_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - column_ = com.google.protobuf.ByteString.EMPTY; + column_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); timestamp_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); - data_ = com.google.protobuf.ByteString.EMPTY; + data_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000008); return this; } @@ -428,7 +428,7 @@ public final class CellMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_descriptor; } @@ -470,7 +470,7 @@ public final class CellMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell)other); } else { @@ -502,13 +502,13 @@ public final class CellMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell) e.getUnfinishedMessage(); throw e; } finally { @@ -521,7 +521,7 @@ public final class CellMessage { private int bitField0_; // optional bytes row = 1; - private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes row = 1; * @@ -539,7 +539,7 @@ public final class CellMessage { * unused if Cell is in a CellSet * */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } /** @@ -549,7 +549,7 @@ public final class CellMessage { * unused if Cell is in a CellSet * */ - public Builder setRow(com.google.protobuf.ByteString value) { + public Builder setRow(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -573,7 +573,7 @@ public final class CellMessage { } // optional bytes column = 2; - private com.google.protobuf.ByteString column_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString column_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes column = 2; */ @@ -583,13 +583,13 @@ public final class CellMessage { /** * optional bytes column = 2; */ - public com.google.protobuf.ByteString getColumn() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getColumn() { return column_; } /** * optional bytes column = 2; */ - public Builder setColumn(com.google.protobuf.ByteString value) { + public Builder setColumn(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -642,7 +642,7 @@ public final class CellMessage { } // optional bytes data = 4; - private com.google.protobuf.ByteString data_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString data_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes data = 4; */ @@ -652,13 +652,13 @@ public final class CellMessage { /** * optional bytes data = 4; */ - public com.google.protobuf.ByteString getData() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getData() { return data_; } /** * optional bytes data = 4; */ - public Builder setData(com.google.protobuf.ByteString value) { + public Builder setData(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -688,17 +688,17 @@ public final class CellMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.Cell) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -707,23 +707,23 @@ public final class CellMessage { "row\030\001 \001(\014\022\016\n\006column\030\002 \001(\014\022\021\n\ttimestamp\030\003" + " \001(\003\022\014\n\004data\030\004 \001(\014" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Cell_descriptor, new java.lang.String[] { "Row", "Column", "Timestamp", "Data", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellSetMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellSetMessage.java index f35a25f..6b962da 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellSetMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/CellSetMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class CellSetMessage { private CellSetMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface CellSetOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.CellSet.Row rows = 1; /** @@ -40,14 +40,14 @@ public final class CellSetMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.CellSet} */ public static final class CellSet extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements CellSetOrBuilder { // Use CellSet.newBuilder() to construct. - private CellSet(com.google.protobuf.GeneratedMessage.Builder builder) { + private CellSet(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private CellSet(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private CellSet(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final CellSet defaultInstance; public static CellSet getDefaultInstance() { @@ -58,20 +58,20 @@ public final class CellSetMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CellSet( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -97,10 +97,10 @@ public final class CellSetMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { @@ -110,35 +110,35 @@ public final class CellSetMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.class, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public CellSet parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new CellSet(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } public interface RowOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes key = 1; /** @@ -148,7 +148,7 @@ public final class CellSetMessage { /** * required bytes key = 1; */ - com.google.protobuf.ByteString getKey(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getKey(); // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.Cell values = 2; /** @@ -179,14 +179,14 @@ public final class CellSetMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.CellSet.Row} */ public static final class Row extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements RowOrBuilder { // Use Row.newBuilder() to construct. - private Row(com.google.protobuf.GeneratedMessage.Builder builder) { + private Row(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Row(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Row(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Row defaultInstance; public static Row getDefaultInstance() { @@ -197,20 +197,20 @@ public final class CellSetMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Row( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -241,10 +241,10 @@ public final class CellSetMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { @@ -254,37 +254,37 @@ public final class CellSetMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row.class, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Row parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Row(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes key = 1; public static final int KEY_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString key_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString key_; /** * required bytes key = 1; */ @@ -294,7 +294,7 @@ public final class CellSetMessage { /** * required bytes key = 1; */ - public com.google.protobuf.ByteString getKey() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getKey() { return key_; } @@ -335,7 +335,7 @@ public final class CellSetMessage { } private void initFields() { - key_ = com.google.protobuf.ByteString.EMPTY; + key_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; values_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; @@ -351,7 +351,7 @@ public final class CellSetMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -370,11 +370,11 @@ public final class CellSetMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, key_); } for (int i = 0; i < values_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(2, values_.get(i)); } size += getUnknownFields().getSerializedSize(); @@ -390,24 +390,24 @@ public final class CellSetMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom(java.io.InputStream input) @@ -416,7 +416,7 @@ public final class CellSetMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -426,18 +426,18 @@ public final class CellSetMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -451,7 +451,7 @@ public final class CellSetMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -459,14 +459,14 @@ public final class CellSetMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.CellSet.Row} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.RowOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -479,12 +479,12 @@ public final class CellSetMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getValuesFieldBuilder(); } } @@ -494,7 +494,7 @@ public final class CellSetMessage { public Builder clear() { super.clear(); - key_ = com.google.protobuf.ByteString.EMPTY; + key_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); if (valuesBuilder_ == null) { values_ = java.util.Collections.emptyList(); @@ -509,7 +509,7 @@ public final class CellSetMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_descriptor; } @@ -548,7 +548,7 @@ public final class CellSetMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row)other); } else { @@ -581,7 +581,7 @@ public final class CellSetMessage { values_ = other.values_; bitField0_ = (bitField0_ & ~0x00000002); valuesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getValuesFieldBuilder() : null; } else { valuesBuilder_.addAllMessages(other.values_); @@ -601,13 +601,13 @@ public final class CellSetMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row) e.getUnfinishedMessage(); throw e; } finally { @@ -620,7 +620,7 @@ public final class CellSetMessage { private int bitField0_; // required bytes key = 1; - private com.google.protobuf.ByteString key_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString key_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes key = 1; */ @@ -630,13 +630,13 @@ public final class CellSetMessage { /** * required bytes key = 1; */ - public com.google.protobuf.ByteString getKey() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getKey() { return key_; } /** * required bytes key = 1; */ - public Builder setKey(com.google.protobuf.ByteString value) { + public Builder setKey(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -665,7 +665,7 @@ public final class CellSetMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.CellOrBuilder> valuesBuilder_; /** @@ -880,11 +880,11 @@ public final class CellSetMessage { getValuesBuilderList() { return getValuesFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.CellOrBuilder> getValuesFieldBuilder() { if (valuesBuilder_ == null) { - valuesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + valuesBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.CellOrBuilder>( values_, ((bitField0_ & 0x00000002) == 0x00000002), @@ -960,7 +960,7 @@ public final class CellSetMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (int i = 0; i < rows_.size(); i++) { @@ -976,7 +976,7 @@ public final class CellSetMessage { size = 0; for (int i = 0; i < rows_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(1, rows_.get(i)); } size += getUnknownFields().getSerializedSize(); @@ -992,24 +992,24 @@ public final class CellSetMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom(java.io.InputStream input) @@ -1018,7 +1018,7 @@ public final class CellSetMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1028,18 +1028,18 @@ public final class CellSetMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1053,7 +1053,7 @@ public final class CellSetMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1061,14 +1061,14 @@ public final class CellSetMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.CellSet} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSetOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1081,12 +1081,12 @@ public final class CellSetMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getRowsFieldBuilder(); } } @@ -1109,7 +1109,7 @@ public final class CellSetMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor; } @@ -1142,7 +1142,7 @@ public final class CellSetMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet)other); } else { @@ -1172,7 +1172,7 @@ public final class CellSetMessage { rows_ = other.rows_; bitField0_ = (bitField0_ & ~0x00000001); rowsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getRowsFieldBuilder() : null; } else { rowsBuilder_.addAllMessages(other.rows_); @@ -1194,13 +1194,13 @@ public final class CellSetMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet) e.getUnfinishedMessage(); throw e; } finally { @@ -1222,7 +1222,7 @@ public final class CellSetMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.RowOrBuilder> rowsBuilder_; /** @@ -1437,11 +1437,11 @@ public final class CellSetMessage { getRowsBuilderList() { return getRowsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.RowOrBuilder> getRowsFieldBuilder() { if (rowsBuilder_ == null) { - rowsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + rowsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.Row.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.CellSetMessage.CellSet.RowOrBuilder>( rows_, ((bitField0_ & 0x00000001) == 0x00000001), @@ -1463,22 +1463,22 @@ public final class CellSetMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.CellSet) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1490,29 +1490,29 @@ public final class CellSetMessage { "\n\006values\030\002 \003(\01325.org.apache.hadoop.hbase" + ".rest.protobuf.generated.Cell" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor, new java.lang.String[] { "Rows", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_CellSet_Row_descriptor, new java.lang.String[] { "Key", "Values", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.getDescriptor(), }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ColumnSchemaMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ColumnSchemaMessage.java index f5f6a95..b560afe 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ColumnSchemaMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ColumnSchemaMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class ColumnSchemaMessage { private ColumnSchemaMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface ColumnSchemaOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional string name = 1; /** @@ -23,7 +23,7 @@ public final class ColumnSchemaMessage { /** * optional string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema.Attribute attrs = 2; @@ -91,21 +91,21 @@ public final class ColumnSchemaMessage { /** * optional string compression = 5; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getCompressionBytes(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema} */ public static final class ColumnSchema extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements ColumnSchemaOrBuilder { // Use ColumnSchema.newBuilder() to construct. - private ColumnSchema(com.google.protobuf.GeneratedMessage.Builder builder) { + private ColumnSchema(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private ColumnSchema(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private ColumnSchema(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final ColumnSchema defaultInstance; public static ColumnSchema getDefaultInstance() { @@ -116,20 +116,20 @@ public final class ColumnSchemaMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ColumnSchema( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -175,10 +175,10 @@ public final class ColumnSchemaMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { @@ -188,35 +188,35 @@ public final class ColumnSchemaMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.class, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public ColumnSchema parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new ColumnSchema(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } public interface AttributeOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string name = 1; /** @@ -230,7 +230,7 @@ public final class ColumnSchemaMessage { /** * required string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // required string value = 2; @@ -245,21 +245,21 @@ public final class ColumnSchemaMessage { /** * required string value = 2; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema.Attribute} */ public static final class Attribute extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements AttributeOrBuilder { // Use Attribute.newBuilder() to construct. - private Attribute(com.google.protobuf.GeneratedMessage.Builder builder) { + private Attribute(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Attribute(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Attribute(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Attribute defaultInstance; public static Attribute getDefaultInstance() { @@ -270,20 +270,20 @@ public final class ColumnSchemaMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Attribute( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -311,40 +311,40 @@ public final class ColumnSchemaMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute.class, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Attribute parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Attribute(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -366,8 +366,8 @@ public final class ColumnSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -378,17 +378,17 @@ public final class ColumnSchemaMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -409,8 +409,8 @@ public final class ColumnSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { value_ = s; @@ -421,17 +421,17 @@ public final class ColumnSchemaMessage { /** * required string value = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -456,7 +456,7 @@ public final class ColumnSchemaMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -475,11 +475,11 @@ public final class ColumnSchemaMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, getValueBytes()); } size += getUnknownFields().getSerializedSize(); @@ -495,24 +495,24 @@ public final class ColumnSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom(java.io.InputStream input) @@ -521,7 +521,7 @@ public final class ColumnSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -531,18 +531,18 @@ public final class ColumnSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -556,7 +556,7 @@ public final class ColumnSchemaMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -564,14 +564,14 @@ public final class ColumnSchemaMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema.Attribute} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.AttributeOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -584,12 +584,12 @@ public final class ColumnSchemaMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -609,7 +609,7 @@ public final class ColumnSchemaMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_descriptor; } @@ -643,7 +643,7 @@ public final class ColumnSchemaMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute)other); } else { @@ -681,13 +681,13 @@ public final class ColumnSchemaMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute) e.getUnfinishedMessage(); throw e; } finally { @@ -713,7 +713,7 @@ public final class ColumnSchemaMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -724,17 +724,17 @@ public final class ColumnSchemaMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -763,7 +763,7 @@ public final class ColumnSchemaMessage { * required string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -787,7 +787,7 @@ public final class ColumnSchemaMessage { public java.lang.String getValue() { java.lang.Object ref = value_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); value_ = s; return s; @@ -798,17 +798,17 @@ public final class ColumnSchemaMessage { /** * required string value = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -837,7 +837,7 @@ public final class ColumnSchemaMessage { * required string value = 2; */ public Builder setValueBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -876,8 +876,8 @@ public final class ColumnSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -888,17 +888,17 @@ public final class ColumnSchemaMessage { /** * optional string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -995,8 +995,8 @@ public final class ColumnSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { compression_ = s; @@ -1007,17 +1007,17 @@ public final class ColumnSchemaMessage { /** * optional string compression = 5; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getCompressionBytes() { java.lang.Object ref = compression_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); compression_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -1043,7 +1043,7 @@ public final class ColumnSchemaMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1071,23 +1071,23 @@ public final class ColumnSchemaMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } for (int i = 0; i < attrs_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(2, attrs_.get(i)); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(3, ttl_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(4, maxVersions_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(5, getCompressionBytes()); } size += getUnknownFields().getSerializedSize(); @@ -1103,24 +1103,24 @@ public final class ColumnSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom(java.io.InputStream input) @@ -1129,7 +1129,7 @@ public final class ColumnSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1139,18 +1139,18 @@ public final class ColumnSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1164,7 +1164,7 @@ public final class ColumnSchemaMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1172,14 +1172,14 @@ public final class ColumnSchemaMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchemaOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1192,12 +1192,12 @@ public final class ColumnSchemaMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getAttrsFieldBuilder(); } } @@ -1228,7 +1228,7 @@ public final class ColumnSchemaMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor; } @@ -1279,7 +1279,7 @@ public final class ColumnSchemaMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema)other); } else { @@ -1314,7 +1314,7 @@ public final class ColumnSchemaMessage { attrs_ = other.attrs_; bitField0_ = (bitField0_ & ~0x00000002); attrsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getAttrsFieldBuilder() : null; } else { attrsBuilder_.addAllMessages(other.attrs_); @@ -1347,13 +1347,13 @@ public final class ColumnSchemaMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema) e.getUnfinishedMessage(); throw e; } finally { @@ -1379,7 +1379,7 @@ public final class ColumnSchemaMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -1390,17 +1390,17 @@ public final class ColumnSchemaMessage { /** * optional string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1429,7 +1429,7 @@ public final class ColumnSchemaMessage { * optional string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1449,7 +1449,7 @@ public final class ColumnSchemaMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.AttributeOrBuilder> attrsBuilder_; /** @@ -1664,11 +1664,11 @@ public final class ColumnSchemaMessage { getAttrsBuilderList() { return getAttrsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.AttributeOrBuilder> getAttrsFieldBuilder() { if (attrsBuilder_ == null) { - attrsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + attrsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Attribute.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.AttributeOrBuilder>( attrs_, ((bitField0_ & 0x00000002) == 0x00000002), @@ -1775,7 +1775,7 @@ public final class ColumnSchemaMessage { public java.lang.String getCompression() { java.lang.Object ref = compression_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); compression_ = s; return s; @@ -1786,17 +1786,17 @@ public final class ColumnSchemaMessage { /** * optional string compression = 5; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getCompressionBytes() { java.lang.Object ref = compression_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); compression_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1825,7 +1825,7 @@ public final class ColumnSchemaMessage { * optional string compression = 5; */ public Builder setCompressionBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1846,22 +1846,22 @@ public final class ColumnSchemaMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1874,29 +1874,29 @@ public final class ColumnSchemaMessage { "sion\030\005 \001(\t\032(\n\tAttribute\022\014\n\004name\030\001 \002(\t\022\r\n" + "\005value\030\002 \002(\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor, new java.lang.String[] { "Name", "Attrs", "Ttl", "MaxVersions", "Compression", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_ColumnSchema_Attribute_descriptor, new java.lang.String[] { "Name", "Value", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacePropertiesMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacePropertiesMessage.java index 3b5a13a..45a7e82 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacePropertiesMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacePropertiesMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class NamespacePropertiesMessage { private NamespacePropertiesMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface NamespacePropertiesOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.NamespaceProperties.Property props = 1; /** @@ -40,14 +40,14 @@ public final class NamespacePropertiesMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.NamespaceProperties} */ public static final class NamespaceProperties extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements NamespacePropertiesOrBuilder { // Use NamespaceProperties.newBuilder() to construct. - private NamespaceProperties(com.google.protobuf.GeneratedMessage.Builder builder) { + private NamespaceProperties(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private NamespaceProperties(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private NamespaceProperties(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final NamespaceProperties defaultInstance; public static NamespaceProperties getDefaultInstance() { @@ -58,20 +58,20 @@ public final class NamespacePropertiesMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private NamespaceProperties( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -97,10 +97,10 @@ public final class NamespacePropertiesMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { @@ -110,35 +110,35 @@ public final class NamespacePropertiesMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.class, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public NamespaceProperties parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new NamespaceProperties(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } public interface PropertyOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string key = 1; /** @@ -152,7 +152,7 @@ public final class NamespacePropertiesMessage { /** * required string key = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getKeyBytes(); // required string value = 2; @@ -167,21 +167,21 @@ public final class NamespacePropertiesMessage { /** * required string value = 2; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.NamespaceProperties.Property} */ public static final class Property extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements PropertyOrBuilder { // Use Property.newBuilder() to construct. - private Property(com.google.protobuf.GeneratedMessage.Builder builder) { + private Property(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Property(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Property(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Property defaultInstance; public static Property getDefaultInstance() { @@ -192,20 +192,20 @@ public final class NamespacePropertiesMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Property( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -233,40 +233,40 @@ public final class NamespacePropertiesMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property.class, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Property parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Property(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -288,8 +288,8 @@ public final class NamespacePropertiesMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { key_ = s; @@ -300,17 +300,17 @@ public final class NamespacePropertiesMessage { /** * required string key = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getKeyBytes() { java.lang.Object ref = key_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); key_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -331,8 +331,8 @@ public final class NamespacePropertiesMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { value_ = s; @@ -343,17 +343,17 @@ public final class NamespacePropertiesMessage { /** * required string value = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -378,7 +378,7 @@ public final class NamespacePropertiesMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -397,11 +397,11 @@ public final class NamespacePropertiesMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getKeyBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, getValueBytes()); } size += getUnknownFields().getSerializedSize(); @@ -417,24 +417,24 @@ public final class NamespacePropertiesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom(java.io.InputStream input) @@ -443,7 +443,7 @@ public final class NamespacePropertiesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -453,18 +453,18 @@ public final class NamespacePropertiesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -478,7 +478,7 @@ public final class NamespacePropertiesMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -486,14 +486,14 @@ public final class NamespacePropertiesMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.NamespaceProperties.Property} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.PropertyOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -506,12 +506,12 @@ public final class NamespacePropertiesMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -531,7 +531,7 @@ public final class NamespacePropertiesMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_descriptor; } @@ -565,7 +565,7 @@ public final class NamespacePropertiesMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property)other); } else { @@ -603,13 +603,13 @@ public final class NamespacePropertiesMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property) e.getUnfinishedMessage(); throw e; } finally { @@ -635,7 +635,7 @@ public final class NamespacePropertiesMessage { public java.lang.String getKey() { java.lang.Object ref = key_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); key_ = s; return s; @@ -646,17 +646,17 @@ public final class NamespacePropertiesMessage { /** * required string key = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getKeyBytes() { java.lang.Object ref = key_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); key_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -685,7 +685,7 @@ public final class NamespacePropertiesMessage { * required string key = 1; */ public Builder setKeyBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -709,7 +709,7 @@ public final class NamespacePropertiesMessage { public java.lang.String getValue() { java.lang.Object ref = value_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); value_ = s; return s; @@ -720,17 +720,17 @@ public final class NamespacePropertiesMessage { /** * required string value = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -759,7 +759,7 @@ public final class NamespacePropertiesMessage { * required string value = 2; */ public Builder setValueBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -834,7 +834,7 @@ public final class NamespacePropertiesMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (int i = 0; i < props_.size(); i++) { @@ -850,7 +850,7 @@ public final class NamespacePropertiesMessage { size = 0; for (int i = 0; i < props_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(1, props_.get(i)); } size += getUnknownFields().getSerializedSize(); @@ -866,24 +866,24 @@ public final class NamespacePropertiesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom(java.io.InputStream input) @@ -892,7 +892,7 @@ public final class NamespacePropertiesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -902,18 +902,18 @@ public final class NamespacePropertiesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -927,7 +927,7 @@ public final class NamespacePropertiesMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -935,14 +935,14 @@ public final class NamespacePropertiesMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.NamespaceProperties} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespacePropertiesOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -955,12 +955,12 @@ public final class NamespacePropertiesMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getPropsFieldBuilder(); } } @@ -983,7 +983,7 @@ public final class NamespacePropertiesMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor; } @@ -1016,7 +1016,7 @@ public final class NamespacePropertiesMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties)other); } else { @@ -1046,7 +1046,7 @@ public final class NamespacePropertiesMessage { props_ = other.props_; bitField0_ = (bitField0_ & ~0x00000001); propsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getPropsFieldBuilder() : null; } else { propsBuilder_.addAllMessages(other.props_); @@ -1068,13 +1068,13 @@ public final class NamespacePropertiesMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties) e.getUnfinishedMessage(); throw e; } finally { @@ -1096,7 +1096,7 @@ public final class NamespacePropertiesMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.PropertyOrBuilder> propsBuilder_; /** @@ -1311,11 +1311,11 @@ public final class NamespacePropertiesMessage { getPropsBuilderList() { return getPropsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.PropertyOrBuilder> getPropsFieldBuilder() { if (propsBuilder_ == null) { - propsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + propsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.Property.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacePropertiesMessage.NamespaceProperties.PropertyOrBuilder>( props_, ((bitField0_ & 0x00000001) == 0x00000001), @@ -1337,22 +1337,22 @@ public final class NamespacePropertiesMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.NamespaceProperties) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1364,29 +1364,29 @@ public final class NamespacePropertiesMessage { "y\032&\n\010Property\022\013\n\003key\030\001 \002(\t\022\r\n\005value\030\002 \002(" + "\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor, new java.lang.String[] { "Props", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_NamespaceProperties_Property_descriptor, new java.lang.String[] { "Key", "Value", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacesMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacesMessage.java index 20725d2..b450361 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacesMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/NamespacesMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class NamespacesMessage { private NamespacesMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface NamespacesOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // repeated string namespace = 1; /** @@ -28,21 +28,21 @@ public final class NamespacesMessage { /** * repeated string namespace = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNamespaceBytes(int index); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Namespaces} */ public static final class Namespaces extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements NamespacesOrBuilder { // Use Namespaces.newBuilder() to construct. - private Namespaces(com.google.protobuf.GeneratedMessage.Builder builder) { + private Namespaces(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Namespaces(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Namespaces(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Namespaces defaultInstance; public static Namespaces getDefaultInstance() { @@ -53,20 +53,20 @@ public final class NamespacesMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Namespaces( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -84,7 +84,7 @@ public final class NamespacesMessage { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - namespace_ = new com.google.protobuf.LazyStringArrayList(); + namespace_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000001; } namespace_.add(input.readBytes()); @@ -92,49 +92,49 @@ public final class NamespacesMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - namespace_ = new com.google.protobuf.UnmodifiableLazyStringList(namespace_); + namespace_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList(namespace_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces.class, org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Namespaces parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Namespaces(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } // repeated string namespace = 1; public static final int NAMESPACE_FIELD_NUMBER = 1; - private com.google.protobuf.LazyStringList namespace_; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList namespace_; /** * repeated string namespace = 1; */ @@ -157,13 +157,13 @@ public final class NamespacesMessage { /** * repeated string namespace = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNamespaceBytes(int index) { return namespace_.getByteString(index); } private void initFields() { - namespace_ = com.google.protobuf.LazyStringArrayList.EMPTY; + namespace_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -174,7 +174,7 @@ public final class NamespacesMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (int i = 0; i < namespace_.size(); i++) { @@ -192,7 +192,7 @@ public final class NamespacesMessage { { int dataSize = 0; for (int i = 0; i < namespace_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(namespace_.getByteString(i)); } size += dataSize; @@ -211,24 +211,24 @@ public final class NamespacesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom(java.io.InputStream input) @@ -237,7 +237,7 @@ public final class NamespacesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -247,18 +247,18 @@ public final class NamespacesMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -272,7 +272,7 @@ public final class NamespacesMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -280,14 +280,14 @@ public final class NamespacesMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Namespaces} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.NamespacesOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -300,12 +300,12 @@ public final class NamespacesMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -314,7 +314,7 @@ public final class NamespacesMessage { public Builder clear() { super.clear(); - namespace_ = com.google.protobuf.LazyStringArrayList.EMPTY; + namespace_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); return this; } @@ -323,7 +323,7 @@ public final class NamespacesMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_descriptor; } @@ -344,7 +344,7 @@ public final class NamespacesMessage { org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces result = new org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) == 0x00000001)) { - namespace_ = new com.google.protobuf.UnmodifiableLazyStringList( + namespace_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList( namespace_); bitField0_ = (bitField0_ & ~0x00000001); } @@ -353,7 +353,7 @@ public final class NamespacesMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces)other); } else { @@ -383,13 +383,13 @@ public final class NamespacesMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.NamespacesMessage.Namespaces) e.getUnfinishedMessage(); throw e; } finally { @@ -402,10 +402,10 @@ public final class NamespacesMessage { private int bitField0_; // repeated string namespace = 1; - private com.google.protobuf.LazyStringList namespace_ = com.google.protobuf.LazyStringArrayList.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList namespace_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureNamespaceIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - namespace_ = new com.google.protobuf.LazyStringArrayList(namespace_); + namespace_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(namespace_); bitField0_ |= 0x00000001; } } @@ -431,7 +431,7 @@ public final class NamespacesMessage { /** * repeated string namespace = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNamespaceBytes(int index) { return namespace_.getByteString(index); } @@ -475,7 +475,7 @@ public final class NamespacesMessage { * repeated string namespace = 1; */ public Builder clearNamespace() { - namespace_ = com.google.protobuf.LazyStringArrayList.EMPTY; + namespace_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; @@ -484,7 +484,7 @@ public final class NamespacesMessage { * repeated string namespace = 1; */ public Builder addNamespaceBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -505,17 +505,17 @@ public final class NamespacesMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.Namespaces) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -523,23 +523,23 @@ public final class NamespacesMessage { "doop.hbase.rest.protobuf.generated\"\037\n\nNa" + "mespaces\022\021\n\tnamespace\030\001 \003(\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Namespaces_descriptor, new java.lang.String[] { "Namespace", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ScannerMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ScannerMessage.java index ef9d1b9..4cba99e 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ScannerMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/ScannerMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class ScannerMessage { private ScannerMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface ScannerOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional bytes startRow = 1; /** @@ -19,7 +19,7 @@ public final class ScannerMessage { /** * optional bytes startRow = 1; */ - com.google.protobuf.ByteString getStartRow(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getStartRow(); // optional bytes endRow = 2; /** @@ -29,7 +29,7 @@ public final class ScannerMessage { /** * optional bytes endRow = 2; */ - com.google.protobuf.ByteString getEndRow(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getEndRow(); // repeated bytes columns = 3; /** @@ -43,7 +43,7 @@ public final class ScannerMessage { /** * repeated bytes columns = 3; */ - com.google.protobuf.ByteString getColumns(int index); + org.apache.hbase.shaded.com.google.protobuf.ByteString getColumns(int index); // optional int32 batch = 4; /** @@ -97,7 +97,7 @@ public final class ScannerMessage { /** * optional string filter = 8; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getFilterBytes(); // optional int32 caching = 9; @@ -135,7 +135,7 @@ public final class ScannerMessage { /** * repeated string labels = 10; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getLabelsBytes(int index); // optional bool cacheBlocks = 11; @@ -160,14 +160,14 @@ public final class ScannerMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Scanner} */ public static final class Scanner extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements ScannerOrBuilder { // Use Scanner.newBuilder() to construct. - private Scanner(com.google.protobuf.GeneratedMessage.Builder builder) { + private Scanner(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Scanner(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Scanner(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Scanner defaultInstance; public static Scanner getDefaultInstance() { @@ -178,20 +178,20 @@ public final class ScannerMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Scanner( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -257,7 +257,7 @@ public final class ScannerMessage { } case 82: { if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) { - labels_ = new com.google.protobuf.LazyStringArrayList(); + labels_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000200; } labels_.add(input.readBytes()); @@ -270,53 +270,53 @@ public final class ScannerMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { columns_ = java.util.Collections.unmodifiableList(columns_); } if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) { - labels_ = new com.google.protobuf.UnmodifiableLazyStringList(labels_); + labels_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList(labels_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner.class, org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Scanner parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Scanner(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // optional bytes startRow = 1; public static final int STARTROW_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString startRow_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString startRow_; /** * optional bytes startRow = 1; */ @@ -326,13 +326,13 @@ public final class ScannerMessage { /** * optional bytes startRow = 1; */ - public com.google.protobuf.ByteString getStartRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getStartRow() { return startRow_; } // optional bytes endRow = 2; public static final int ENDROW_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString endRow_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString endRow_; /** * optional bytes endRow = 2; */ @@ -342,7 +342,7 @@ public final class ScannerMessage { /** * optional bytes endRow = 2; */ - public com.google.protobuf.ByteString getEndRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getEndRow() { return endRow_; } @@ -365,7 +365,7 @@ public final class ScannerMessage { /** * repeated bytes columns = 3; */ - public com.google.protobuf.ByteString getColumns(int index) { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getColumns(int index) { return columns_.get(index); } @@ -450,8 +450,8 @@ public final class ScannerMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { filter_ = s; @@ -462,17 +462,17 @@ public final class ScannerMessage { /** * optional string filter = 8; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFilterBytes() { java.lang.Object ref = filter_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); filter_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -502,7 +502,7 @@ public final class ScannerMessage { // repeated string labels = 10; public static final int LABELS_FIELD_NUMBER = 10; - private com.google.protobuf.LazyStringList labels_; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList labels_; /** * repeated string labels = 10; */ @@ -525,7 +525,7 @@ public final class ScannerMessage { /** * repeated string labels = 10; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getLabelsBytes(int index) { return labels_.getByteString(index); } @@ -555,8 +555,8 @@ public final class ScannerMessage { } private void initFields() { - startRow_ = com.google.protobuf.ByteString.EMPTY; - endRow_ = com.google.protobuf.ByteString.EMPTY; + startRow_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + endRow_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; columns_ = java.util.Collections.emptyList(); batch_ = 0; startTime_ = 0L; @@ -564,7 +564,7 @@ public final class ScannerMessage { maxVersions_ = 0; filter_ = ""; caching_ = 0; - labels_ = com.google.protobuf.LazyStringArrayList.EMPTY; + labels_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; cacheBlocks_ = false; } private byte memoizedIsInitialized = -1; @@ -576,7 +576,7 @@ public final class ScannerMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -622,57 +622,57 @@ public final class ScannerMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, startRow_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, endRow_); } { int dataSize = 0; for (int i = 0; i < columns_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(columns_.get(i)); } size += dataSize; size += 1 * getColumnsList().size(); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(4, batch_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(5, startTime_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(6, endTime_); } if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(7, maxVersions_); } if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(8, getFilterBytes()); } if (((bitField0_ & 0x00000080) == 0x00000080)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(9, caching_); } { int dataSize = 0; for (int i = 0; i < labels_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(labels_.getByteString(i)); } size += dataSize; size += 1 * getLabelsList().size(); } if (((bitField0_ & 0x00000100) == 0x00000100)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBoolSize(11, cacheBlocks_); } size += getUnknownFields().getSerializedSize(); @@ -688,24 +688,24 @@ public final class ScannerMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom(java.io.InputStream input) @@ -714,7 +714,7 @@ public final class ScannerMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -724,18 +724,18 @@ public final class ScannerMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -749,7 +749,7 @@ public final class ScannerMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -757,14 +757,14 @@ public final class ScannerMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Scanner} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.ScannerOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -777,12 +777,12 @@ public final class ScannerMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -791,9 +791,9 @@ public final class ScannerMessage { public Builder clear() { super.clear(); - startRow_ = com.google.protobuf.ByteString.EMPTY; + startRow_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - endRow_ = com.google.protobuf.ByteString.EMPTY; + endRow_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); columns_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); @@ -809,7 +809,7 @@ public final class ScannerMessage { bitField0_ = (bitField0_ & ~0x00000080); caching_ = 0; bitField0_ = (bitField0_ & ~0x00000100); - labels_ = com.google.protobuf.LazyStringArrayList.EMPTY; + labels_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000200); cacheBlocks_ = false; bitField0_ = (bitField0_ & ~0x00000400); @@ -820,7 +820,7 @@ public final class ScannerMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_descriptor; } @@ -879,7 +879,7 @@ public final class ScannerMessage { } result.caching_ = caching_; if (((bitField0_ & 0x00000200) == 0x00000200)) { - labels_ = new com.google.protobuf.UnmodifiableLazyStringList( + labels_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList( labels_); bitField0_ = (bitField0_ & ~0x00000200); } @@ -893,7 +893,7 @@ public final class ScannerMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner)other); } else { @@ -962,13 +962,13 @@ public final class ScannerMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.ScannerMessage.Scanner) e.getUnfinishedMessage(); throw e; } finally { @@ -981,7 +981,7 @@ public final class ScannerMessage { private int bitField0_; // optional bytes startRow = 1; - private com.google.protobuf.ByteString startRow_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString startRow_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes startRow = 1; */ @@ -991,13 +991,13 @@ public final class ScannerMessage { /** * optional bytes startRow = 1; */ - public com.google.protobuf.ByteString getStartRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getStartRow() { return startRow_; } /** * optional bytes startRow = 1; */ - public Builder setStartRow(com.google.protobuf.ByteString value) { + public Builder setStartRow(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1017,7 +1017,7 @@ public final class ScannerMessage { } // optional bytes endRow = 2; - private com.google.protobuf.ByteString endRow_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString endRow_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes endRow = 2; */ @@ -1027,13 +1027,13 @@ public final class ScannerMessage { /** * optional bytes endRow = 2; */ - public com.google.protobuf.ByteString getEndRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getEndRow() { return endRow_; } /** * optional bytes endRow = 2; */ - public Builder setEndRow(com.google.protobuf.ByteString value) { + public Builder setEndRow(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1076,14 +1076,14 @@ public final class ScannerMessage { /** * repeated bytes columns = 3; */ - public com.google.protobuf.ByteString getColumns(int index) { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getColumns(int index) { return columns_.get(index); } /** * repeated bytes columns = 3; */ public Builder setColumns( - int index, com.google.protobuf.ByteString value) { + int index, org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1095,7 +1095,7 @@ public final class ScannerMessage { /** * repeated bytes columns = 3; */ - public Builder addColumns(com.google.protobuf.ByteString value) { + public Builder addColumns(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1108,7 +1108,7 @@ public final class ScannerMessage { * repeated bytes columns = 3; */ public Builder addAllColumns( - java.lang.Iterable values) { + java.lang.Iterable values) { ensureColumnsIsMutable(); super.addAll(values, columns_); onChanged(); @@ -1270,7 +1270,7 @@ public final class ScannerMessage { public java.lang.String getFilter() { java.lang.Object ref = filter_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); filter_ = s; return s; @@ -1281,17 +1281,17 @@ public final class ScannerMessage { /** * optional string filter = 8; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFilterBytes() { java.lang.Object ref = filter_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); filter_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1320,7 +1320,7 @@ public final class ScannerMessage { * optional string filter = 8; */ public Builder setFilterBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1380,10 +1380,10 @@ public final class ScannerMessage { } // repeated string labels = 10; - private com.google.protobuf.LazyStringList labels_ = com.google.protobuf.LazyStringArrayList.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList labels_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureLabelsIsMutable() { if (!((bitField0_ & 0x00000200) == 0x00000200)) { - labels_ = new com.google.protobuf.LazyStringArrayList(labels_); + labels_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(labels_); bitField0_ |= 0x00000200; } } @@ -1409,7 +1409,7 @@ public final class ScannerMessage { /** * repeated string labels = 10; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getLabelsBytes(int index) { return labels_.getByteString(index); } @@ -1453,7 +1453,7 @@ public final class ScannerMessage { * repeated string labels = 10; */ public Builder clearLabels() { - labels_ = com.google.protobuf.LazyStringArrayList.EMPTY; + labels_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000200); onChanged(); return this; @@ -1462,7 +1462,7 @@ public final class ScannerMessage { * repeated string labels = 10; */ public Builder addLabelsBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1532,17 +1532,17 @@ public final class ScannerMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.Scanner) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1554,23 +1554,23 @@ public final class ScannerMessage { "\030\007 \001(\005\022\016\n\006filter\030\010 \001(\t\022\017\n\007caching\030\t \001(\005\022" + "\016\n\006labels\030\n \003(\t\022\023\n\013cacheBlocks\030\013 \001(\010" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Scanner_descriptor, new java.lang.String[] { "StartRow", "EndRow", "Columns", "Batch", "StartTime", "EndTime", "MaxVersions", "Filter", "Caching", "Labels", "CacheBlocks", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java index 44dce76..2733953 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class StorageClusterStatusMessage { private StorageClusterStatusMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface StorageClusterStatusOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1; /** @@ -73,7 +73,7 @@ public final class StorageClusterStatusMessage { /** * repeated string deadNodes = 2; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getDeadNodesBytes(int index); // optional int32 regions = 3; @@ -118,14 +118,14 @@ public final class StorageClusterStatusMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus} */ public static final class StorageClusterStatus extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements StorageClusterStatusOrBuilder { // Use StorageClusterStatus.newBuilder() to construct. - private StorageClusterStatus(com.google.protobuf.GeneratedMessage.Builder builder) { + private StorageClusterStatus(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private StorageClusterStatus(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private StorageClusterStatus(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final StorageClusterStatus defaultInstance; public static StorageClusterStatus getDefaultInstance() { @@ -136,20 +136,20 @@ public final class StorageClusterStatusMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private StorageClusterStatus( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -175,7 +175,7 @@ public final class StorageClusterStatusMessage { } case 18: { if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { - deadNodes_ = new com.google.protobuf.LazyStringArrayList(); + deadNodes_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000002; } deadNodes_.add(input.readBytes()); @@ -198,51 +198,51 @@ public final class StorageClusterStatusMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { liveNodes_ = java.util.Collections.unmodifiableList(liveNodes_); } if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { - deadNodes_ = new com.google.protobuf.UnmodifiableLazyStringList(deadNodes_); + deadNodes_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList(deadNodes_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.class, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public StorageClusterStatus parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new StorageClusterStatus(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } public interface RegionOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes name = 1; /** @@ -252,7 +252,7 @@ public final class StorageClusterStatusMessage { /** * required bytes name = 1; */ - com.google.protobuf.ByteString getName(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getName(); // optional int32 stores = 2; /** @@ -378,14 +378,14 @@ public final class StorageClusterStatusMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region} */ public static final class Region extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements RegionOrBuilder { // Use Region.newBuilder() to construct. - private Region(com.google.protobuf.GeneratedMessage.Builder builder) { + private Region(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Region(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Region(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Region defaultInstance; public static Region getDefaultInstance() { @@ -396,20 +396,20 @@ public final class StorageClusterStatusMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Region( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -492,47 +492,47 @@ public final class StorageClusterStatusMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.class, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Region parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Region(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes name = 1; public static final int NAME_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString name_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString name_; /** * required bytes name = 1; */ @@ -542,7 +542,7 @@ public final class StorageClusterStatusMessage { /** * required bytes name = 1; */ - public com.google.protobuf.ByteString getName() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getName() { return name_; } @@ -739,7 +739,7 @@ public final class StorageClusterStatusMessage { } private void initFields() { - name_ = com.google.protobuf.ByteString.EMPTY; + name_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; stores_ = 0; storefiles_ = 0; storefileSizeMB_ = 0; @@ -766,7 +766,7 @@ public final class StorageClusterStatusMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -818,55 +818,55 @@ public final class StorageClusterStatusMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, name_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(2, stores_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(3, storefiles_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(4, storefileSizeMB_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(5, memstoreSizeMB_); } if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(6, storefileIndexSizeMB_); } if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(7, readRequestsCount_); } if (((bitField0_ & 0x00000080) == 0x00000080)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(8, writeRequestsCount_); } if (((bitField0_ & 0x00000100) == 0x00000100)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(9, rootIndexSizeKB_); } if (((bitField0_ & 0x00000200) == 0x00000200)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(10, totalStaticIndexSizeKB_); } if (((bitField0_ & 0x00000400) == 0x00000400)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(11, totalStaticBloomSizeKB_); } if (((bitField0_ & 0x00000800) == 0x00000800)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(12, totalCompactingKVs_); } if (((bitField0_ & 0x00001000) == 0x00001000)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(13, currentCompactedKVs_); } size += getUnknownFields().getSerializedSize(); @@ -882,24 +882,24 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(java.io.InputStream input) @@ -908,7 +908,7 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -918,18 +918,18 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -943,7 +943,7 @@ public final class StorageClusterStatusMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -951,14 +951,14 @@ public final class StorageClusterStatusMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -971,12 +971,12 @@ public final class StorageClusterStatusMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -985,7 +985,7 @@ public final class StorageClusterStatusMessage { public Builder clear() { super.clear(); - name_ = com.google.protobuf.ByteString.EMPTY; + name_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); stores_ = 0; bitField0_ = (bitField0_ & ~0x00000002); @@ -1018,7 +1018,7 @@ public final class StorageClusterStatusMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor; } @@ -1096,7 +1096,7 @@ public final class StorageClusterStatusMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region)other); } else { @@ -1159,13 +1159,13 @@ public final class StorageClusterStatusMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region) e.getUnfinishedMessage(); throw e; } finally { @@ -1178,7 +1178,7 @@ public final class StorageClusterStatusMessage { private int bitField0_; // required bytes name = 1; - private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString name_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes name = 1; */ @@ -1188,13 +1188,13 @@ public final class StorageClusterStatusMessage { /** * required bytes name = 1; */ - public com.google.protobuf.ByteString getName() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getName() { return name_; } /** * required bytes name = 1; */ - public Builder setName(com.google.protobuf.ByteString value) { + public Builder setName(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1621,7 +1621,7 @@ public final class StorageClusterStatusMessage { } public interface NodeOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string name = 1; /** @@ -1647,7 +1647,7 @@ public final class StorageClusterStatusMessage { * name:port * */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // optional int64 startCode = 2; @@ -1719,14 +1719,14 @@ public final class StorageClusterStatusMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node} */ public static final class Node extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements NodeOrBuilder { // Use Node.newBuilder() to construct. - private Node(com.google.protobuf.GeneratedMessage.Builder builder) { + private Node(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Node(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Node(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Node defaultInstance; public static Node getDefaultInstance() { @@ -1737,20 +1737,20 @@ public final class StorageClusterStatusMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Node( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1801,10 +1801,10 @@ public final class StorageClusterStatusMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) { @@ -1814,30 +1814,30 @@ public final class StorageClusterStatusMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.class, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Node parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Node(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -1867,8 +1867,8 @@ public final class StorageClusterStatusMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -1883,17 +1883,17 @@ public final class StorageClusterStatusMessage { * name:port * */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -2024,7 +2024,7 @@ public final class StorageClusterStatusMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -2055,27 +2055,27 @@ public final class StorageClusterStatusMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(2, startCode_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(3, requests_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(4, heapSizeMB_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(5, maxHeapSizeMB_); } for (int i = 0; i < regions_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(6, regions_.get(i)); } size += getUnknownFields().getSerializedSize(); @@ -2091,24 +2091,24 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(java.io.InputStream input) @@ -2117,7 +2117,7 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2127,18 +2127,18 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2152,7 +2152,7 @@ public final class StorageClusterStatusMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2160,14 +2160,14 @@ public final class StorageClusterStatusMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2180,12 +2180,12 @@ public final class StorageClusterStatusMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getRegionsFieldBuilder(); } } @@ -2218,7 +2218,7 @@ public final class StorageClusterStatusMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor; } @@ -2273,7 +2273,7 @@ public final class StorageClusterStatusMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node)other); } else { @@ -2320,7 +2320,7 @@ public final class StorageClusterStatusMessage { regions_ = other.regions_; bitField0_ = (bitField0_ & ~0x00000020); regionsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getRegionsFieldBuilder() : null; } else { regionsBuilder_.addAllMessages(other.regions_); @@ -2346,13 +2346,13 @@ public final class StorageClusterStatusMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node) e.getUnfinishedMessage(); throw e; } finally { @@ -2386,7 +2386,7 @@ public final class StorageClusterStatusMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -2401,17 +2401,17 @@ public final class StorageClusterStatusMessage { * name:port * */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -2452,7 +2452,7 @@ public final class StorageClusterStatusMessage { * */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -2604,7 +2604,7 @@ public final class StorageClusterStatusMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> regionsBuilder_; /** @@ -2819,11 +2819,11 @@ public final class StorageClusterStatusMessage { getRegionsBuilderList() { return getRegionsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> getRegionsFieldBuilder() { if (regionsBuilder_ == null) { - regionsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + regionsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder>( regions_, ((bitField0_ & 0x00000020) == 0x00000020), @@ -2904,7 +2904,7 @@ public final class StorageClusterStatusMessage { // repeated string deadNodes = 2; public static final int DEADNODES_FIELD_NUMBER = 2; - private com.google.protobuf.LazyStringList deadNodes_; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList deadNodes_; /** * repeated string deadNodes = 2; */ @@ -2927,7 +2927,7 @@ public final class StorageClusterStatusMessage { /** * repeated string deadNodes = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getDeadNodesBytes(int index) { return deadNodes_.getByteString(index); } @@ -2990,7 +2990,7 @@ public final class StorageClusterStatusMessage { private void initFields() { liveNodes_ = java.util.Collections.emptyList(); - deadNodes_ = com.google.protobuf.LazyStringArrayList.EMPTY; + deadNodes_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; regions_ = 0; requests_ = 0L; averageLoad_ = 0D; @@ -3010,7 +3010,7 @@ public final class StorageClusterStatusMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (int i = 0; i < liveNodes_.size(); i++) { @@ -3038,28 +3038,28 @@ public final class StorageClusterStatusMessage { size = 0; for (int i = 0; i < liveNodes_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(1, liveNodes_.get(i)); } { int dataSize = 0; for (int i = 0; i < deadNodes_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(deadNodes_.getByteString(i)); } size += dataSize; size += 1 * getDeadNodesList().size(); } if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(3, regions_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(4, requests_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeDoubleSize(5, averageLoad_); } size += getUnknownFields().getSerializedSize(); @@ -3075,24 +3075,24 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(java.io.InputStream input) @@ -3101,7 +3101,7 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3111,18 +3111,18 @@ public final class StorageClusterStatusMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3136,7 +3136,7 @@ public final class StorageClusterStatusMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -3144,14 +3144,14 @@ public final class StorageClusterStatusMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatusOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -3164,12 +3164,12 @@ public final class StorageClusterStatusMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getLiveNodesFieldBuilder(); } } @@ -3185,7 +3185,7 @@ public final class StorageClusterStatusMessage { } else { liveNodesBuilder_.clear(); } - deadNodes_ = com.google.protobuf.LazyStringArrayList.EMPTY; + deadNodes_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); regions_ = 0; bitField0_ = (bitField0_ & ~0x00000004); @@ -3200,7 +3200,7 @@ public final class StorageClusterStatusMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor; } @@ -3231,7 +3231,7 @@ public final class StorageClusterStatusMessage { result.liveNodes_ = liveNodesBuilder_.build(); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - deadNodes_ = new com.google.protobuf.UnmodifiableLazyStringList( + deadNodes_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList( deadNodes_); bitField0_ = (bitField0_ & ~0x00000002); } @@ -3253,7 +3253,7 @@ public final class StorageClusterStatusMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus)other); } else { @@ -3283,7 +3283,7 @@ public final class StorageClusterStatusMessage { liveNodes_ = other.liveNodes_; bitField0_ = (bitField0_ & ~0x00000001); liveNodesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getLiveNodesFieldBuilder() : null; } else { liveNodesBuilder_.addAllMessages(other.liveNodes_); @@ -3324,13 +3324,13 @@ public final class StorageClusterStatusMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus) e.getUnfinishedMessage(); throw e; } finally { @@ -3352,7 +3352,7 @@ public final class StorageClusterStatusMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder> liveNodesBuilder_; /** @@ -3639,11 +3639,11 @@ public final class StorageClusterStatusMessage { getLiveNodesBuilderList() { return getLiveNodesFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder> getLiveNodesFieldBuilder() { if (liveNodesBuilder_ == null) { - liveNodesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + liveNodesBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder>( liveNodes_, ((bitField0_ & 0x00000001) == 0x00000001), @@ -3655,10 +3655,10 @@ public final class StorageClusterStatusMessage { } // repeated string deadNodes = 2; - private com.google.protobuf.LazyStringList deadNodes_ = com.google.protobuf.LazyStringArrayList.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList deadNodes_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureDeadNodesIsMutable() { if (!((bitField0_ & 0x00000002) == 0x00000002)) { - deadNodes_ = new com.google.protobuf.LazyStringArrayList(deadNodes_); + deadNodes_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(deadNodes_); bitField0_ |= 0x00000002; } } @@ -3684,7 +3684,7 @@ public final class StorageClusterStatusMessage { /** * repeated string deadNodes = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getDeadNodesBytes(int index) { return deadNodes_.getByteString(index); } @@ -3728,7 +3728,7 @@ public final class StorageClusterStatusMessage { * repeated string deadNodes = 2; */ public Builder clearDeadNodes() { - deadNodes_ = com.google.protobuf.LazyStringArrayList.EMPTY; + deadNodes_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); onChanged(); return this; @@ -3737,7 +3737,7 @@ public final class StorageClusterStatusMessage { * repeated string deadNodes = 2; */ public Builder addDeadNodesBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -3873,27 +3873,27 @@ public final class StorageClusterStatusMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -3919,35 +3919,35 @@ public final class StorageClusterStatusMessage { ".hbase.rest.protobuf.generated.StorageCl", "usterStatus.Region" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor, new java.lang.String[] { "LiveNodes", "DeadNodes", "Regions", "Requests", "AverageLoad", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor, new java.lang.String[] { "Name", "Stores", "Storefiles", "StorefileSizeMB", "MemstoreSizeMB", "StorefileIndexSizeMB", "ReadRequestsCount", "WriteRequestsCount", "RootIndexSizeKB", "TotalStaticIndexSizeKB", "TotalStaticBloomSizeKB", "TotalCompactingKVs", "CurrentCompactedKVs", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(1); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor, new java.lang.String[] { "Name", "StartCode", "Requests", "HeapSizeMB", "MaxHeapSizeMB", "Regions", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableInfoMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableInfoMessage.java index 421c0ec..52dff35 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableInfoMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableInfoMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class TableInfoMessage { private TableInfoMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface TableInfoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string name = 1; /** @@ -23,7 +23,7 @@ public final class TableInfoMessage { /** * required string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.TableInfo.Region regions = 2; @@ -55,14 +55,14 @@ public final class TableInfoMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableInfo} */ public static final class TableInfo extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements TableInfoOrBuilder { // Use TableInfo.newBuilder() to construct. - private TableInfo(com.google.protobuf.GeneratedMessage.Builder builder) { + private TableInfo(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private TableInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private TableInfo(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TableInfo defaultInstance; public static TableInfo getDefaultInstance() { @@ -73,20 +73,20 @@ public final class TableInfoMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TableInfo( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -117,10 +117,10 @@ public final class TableInfoMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { @@ -130,35 +130,35 @@ public final class TableInfoMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.class, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public TableInfo parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new TableInfo(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } public interface RegionOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string name = 1; /** @@ -172,7 +172,7 @@ public final class TableInfoMessage { /** * required string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // optional bytes startKey = 2; @@ -183,7 +183,7 @@ public final class TableInfoMessage { /** * optional bytes startKey = 2; */ - com.google.protobuf.ByteString getStartKey(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getStartKey(); // optional bytes endKey = 3; /** @@ -193,7 +193,7 @@ public final class TableInfoMessage { /** * optional bytes endKey = 3; */ - com.google.protobuf.ByteString getEndKey(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getEndKey(); // optional int64 id = 4; /** @@ -217,21 +217,21 @@ public final class TableInfoMessage { /** * optional string location = 5; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getLocationBytes(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableInfo.Region} */ public static final class Region extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements RegionOrBuilder { // Use Region.newBuilder() to construct. - private Region(com.google.protobuf.GeneratedMessage.Builder builder) { + private Region(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Region(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Region(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Region defaultInstance; public static Region getDefaultInstance() { @@ -242,20 +242,20 @@ public final class TableInfoMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Region( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -298,40 +298,40 @@ public final class TableInfoMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region.class, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Region parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Region(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -353,8 +353,8 @@ public final class TableInfoMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -365,23 +365,23 @@ public final class TableInfoMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } // optional bytes startKey = 2; public static final int STARTKEY_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString startKey_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString startKey_; /** * optional bytes startKey = 2; */ @@ -391,13 +391,13 @@ public final class TableInfoMessage { /** * optional bytes startKey = 2; */ - public com.google.protobuf.ByteString getStartKey() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getStartKey() { return startKey_; } // optional bytes endKey = 3; public static final int ENDKEY_FIELD_NUMBER = 3; - private com.google.protobuf.ByteString endKey_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString endKey_; /** * optional bytes endKey = 3; */ @@ -407,7 +407,7 @@ public final class TableInfoMessage { /** * optional bytes endKey = 3; */ - public com.google.protobuf.ByteString getEndKey() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getEndKey() { return endKey_; } @@ -444,8 +444,8 @@ public final class TableInfoMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { location_ = s; @@ -456,24 +456,24 @@ public final class TableInfoMessage { /** * optional string location = 5; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getLocationBytes() { java.lang.Object ref = location_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); location_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } private void initFields() { name_ = ""; - startKey_ = com.google.protobuf.ByteString.EMPTY; - endKey_ = com.google.protobuf.ByteString.EMPTY; + startKey_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + endKey_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; id_ = 0L; location_ = ""; } @@ -490,7 +490,7 @@ public final class TableInfoMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -518,23 +518,23 @@ public final class TableInfoMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, startKey_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(3, endKey_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(4, id_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(5, getLocationBytes()); } size += getUnknownFields().getSerializedSize(); @@ -550,24 +550,24 @@ public final class TableInfoMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom(java.io.InputStream input) @@ -576,7 +576,7 @@ public final class TableInfoMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -586,18 +586,18 @@ public final class TableInfoMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -611,7 +611,7 @@ public final class TableInfoMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -619,14 +619,14 @@ public final class TableInfoMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableInfo.Region} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.RegionOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -639,12 +639,12 @@ public final class TableInfoMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -655,9 +655,9 @@ public final class TableInfoMessage { super.clear(); name_ = ""; bitField0_ = (bitField0_ & ~0x00000001); - startKey_ = com.google.protobuf.ByteString.EMPTY; + startKey_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); - endKey_ = com.google.protobuf.ByteString.EMPTY; + endKey_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000004); id_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); @@ -670,7 +670,7 @@ public final class TableInfoMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_descriptor; } @@ -716,7 +716,7 @@ public final class TableInfoMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region)other); } else { @@ -759,13 +759,13 @@ public final class TableInfoMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region) e.getUnfinishedMessage(); throw e; } finally { @@ -791,7 +791,7 @@ public final class TableInfoMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -802,17 +802,17 @@ public final class TableInfoMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -841,7 +841,7 @@ public final class TableInfoMessage { * required string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -852,7 +852,7 @@ public final class TableInfoMessage { } // optional bytes startKey = 2; - private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString startKey_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes startKey = 2; */ @@ -862,13 +862,13 @@ public final class TableInfoMessage { /** * optional bytes startKey = 2; */ - public com.google.protobuf.ByteString getStartKey() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getStartKey() { return startKey_; } /** * optional bytes startKey = 2; */ - public Builder setStartKey(com.google.protobuf.ByteString value) { + public Builder setStartKey(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -888,7 +888,7 @@ public final class TableInfoMessage { } // optional bytes endKey = 3; - private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString endKey_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes endKey = 3; */ @@ -898,13 +898,13 @@ public final class TableInfoMessage { /** * optional bytes endKey = 3; */ - public com.google.protobuf.ByteString getEndKey() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getEndKey() { return endKey_; } /** * optional bytes endKey = 3; */ - public Builder setEndKey(com.google.protobuf.ByteString value) { + public Builder setEndKey(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -970,7 +970,7 @@ public final class TableInfoMessage { public java.lang.String getLocation() { java.lang.Object ref = location_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); location_ = s; return s; @@ -981,17 +981,17 @@ public final class TableInfoMessage { /** * optional string location = 5; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getLocationBytes() { java.lang.Object ref = location_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); location_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1020,7 +1020,7 @@ public final class TableInfoMessage { * optional string location = 5; */ public Builder setLocationBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1059,8 +1059,8 @@ public final class TableInfoMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -1071,17 +1071,17 @@ public final class TableInfoMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -1144,7 +1144,7 @@ public final class TableInfoMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1163,11 +1163,11 @@ public final class TableInfoMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } for (int i = 0; i < regions_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(2, regions_.get(i)); } size += getUnknownFields().getSerializedSize(); @@ -1183,24 +1183,24 @@ public final class TableInfoMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom(java.io.InputStream input) @@ -1209,7 +1209,7 @@ public final class TableInfoMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1219,18 +1219,18 @@ public final class TableInfoMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1244,7 +1244,7 @@ public final class TableInfoMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1252,14 +1252,14 @@ public final class TableInfoMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableInfo} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1272,12 +1272,12 @@ public final class TableInfoMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getRegionsFieldBuilder(); } } @@ -1302,7 +1302,7 @@ public final class TableInfoMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor; } @@ -1341,7 +1341,7 @@ public final class TableInfoMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo)other); } else { @@ -1376,7 +1376,7 @@ public final class TableInfoMessage { regions_ = other.regions_; bitField0_ = (bitField0_ & ~0x00000002); regionsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getRegionsFieldBuilder() : null; } else { regionsBuilder_.addAllMessages(other.regions_); @@ -1402,13 +1402,13 @@ public final class TableInfoMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo) e.getUnfinishedMessage(); throw e; } finally { @@ -1434,7 +1434,7 @@ public final class TableInfoMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -1445,17 +1445,17 @@ public final class TableInfoMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1484,7 +1484,7 @@ public final class TableInfoMessage { * required string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1504,7 +1504,7 @@ public final class TableInfoMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.RegionOrBuilder> regionsBuilder_; /** @@ -1719,11 +1719,11 @@ public final class TableInfoMessage { getRegionsBuilderList() { return getRegionsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.RegionOrBuilder> getRegionsFieldBuilder() { if (regionsBuilder_ == null) { - regionsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + regionsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo.RegionOrBuilder>( regions_, ((bitField0_ & 0x00000002) == 0x00000002), @@ -1745,22 +1745,22 @@ public final class TableInfoMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableInfo) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1772,29 +1772,29 @@ public final class TableInfoMessage { "ame\030\001 \002(\t\022\020\n\010startKey\030\002 \001(\014\022\016\n\006endKey\030\003 " + "\001(\014\022\n\n\002id\030\004 \001(\003\022\020\n\010location\030\005 \001(\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor, new java.lang.String[] { "Name", "Regions", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableInfo_Region_descriptor, new java.lang.String[] { "Name", "StartKey", "EndKey", "Id", "Location", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableListMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableListMessage.java index 20891a2..a794e42 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableListMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableListMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class TableListMessage { private TableListMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface TableListOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // repeated string name = 1; /** @@ -28,21 +28,21 @@ public final class TableListMessage { /** * repeated string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(int index); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableList} */ public static final class TableList extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements TableListOrBuilder { // Use TableList.newBuilder() to construct. - private TableList(com.google.protobuf.GeneratedMessage.Builder builder) { + private TableList(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private TableList(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private TableList(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TableList defaultInstance; public static TableList getDefaultInstance() { @@ -53,20 +53,20 @@ public final class TableListMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TableList( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -84,7 +84,7 @@ public final class TableListMessage { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - name_ = new com.google.protobuf.LazyStringArrayList(); + name_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000001; } name_.add(input.readBytes()); @@ -92,49 +92,49 @@ public final class TableListMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - name_ = new com.google.protobuf.UnmodifiableLazyStringList(name_); + name_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList(name_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList.class, org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public TableList parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new TableList(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } // repeated string name = 1; public static final int NAME_FIELD_NUMBER = 1; - private com.google.protobuf.LazyStringList name_; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList name_; /** * repeated string name = 1; */ @@ -157,13 +157,13 @@ public final class TableListMessage { /** * repeated string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(int index) { return name_.getByteString(index); } private void initFields() { - name_ = com.google.protobuf.LazyStringArrayList.EMPTY; + name_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -174,7 +174,7 @@ public final class TableListMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (int i = 0; i < name_.size(); i++) { @@ -192,7 +192,7 @@ public final class TableListMessage { { int dataSize = 0; for (int i = 0; i < name_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(name_.getByteString(i)); } size += dataSize; @@ -211,24 +211,24 @@ public final class TableListMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom(java.io.InputStream input) @@ -237,7 +237,7 @@ public final class TableListMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -247,18 +247,18 @@ public final class TableListMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -272,7 +272,7 @@ public final class TableListMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -280,14 +280,14 @@ public final class TableListMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableList} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableListOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -300,12 +300,12 @@ public final class TableListMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -314,7 +314,7 @@ public final class TableListMessage { public Builder clear() { super.clear(); - name_ = com.google.protobuf.LazyStringArrayList.EMPTY; + name_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); return this; } @@ -323,7 +323,7 @@ public final class TableListMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_descriptor; } @@ -344,7 +344,7 @@ public final class TableListMessage { org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList result = new org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) == 0x00000001)) { - name_ = new com.google.protobuf.UnmodifiableLazyStringList( + name_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList( name_); bitField0_ = (bitField0_ & ~0x00000001); } @@ -353,7 +353,7 @@ public final class TableListMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList)other); } else { @@ -383,13 +383,13 @@ public final class TableListMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList) e.getUnfinishedMessage(); throw e; } finally { @@ -402,10 +402,10 @@ public final class TableListMessage { private int bitField0_; // repeated string name = 1; - private com.google.protobuf.LazyStringList name_ = com.google.protobuf.LazyStringArrayList.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList name_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureNameIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - name_ = new com.google.protobuf.LazyStringArrayList(name_); + name_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(name_); bitField0_ |= 0x00000001; } } @@ -431,7 +431,7 @@ public final class TableListMessage { /** * repeated string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(int index) { return name_.getByteString(index); } @@ -475,7 +475,7 @@ public final class TableListMessage { * repeated string name = 1; */ public Builder clearName() { - name_ = com.google.protobuf.LazyStringArrayList.EMPTY; + name_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; @@ -484,7 +484,7 @@ public final class TableListMessage { * repeated string name = 1; */ public Builder addNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -505,17 +505,17 @@ public final class TableListMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableList) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -523,23 +523,23 @@ public final class TableListMessage { "oop.hbase.rest.protobuf.generated\"\031\n\tTab" + "leList\022\014\n\004name\030\001 \003(\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableList_descriptor, new java.lang.String[] { "Name", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java index 88f9cd3..6ff9223 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class TableSchemaMessage { private TableSchemaMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface TableSchemaOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional string name = 1; /** @@ -23,7 +23,7 @@ public final class TableSchemaMessage { /** * optional string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute attrs = 2; @@ -108,14 +108,14 @@ public final class TableSchemaMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema} */ public static final class TableSchema extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements TableSchemaOrBuilder { // Use TableSchema.newBuilder() to construct. - private TableSchema(com.google.protobuf.GeneratedMessage.Builder builder) { + private TableSchema(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private TableSchema(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private TableSchema(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TableSchema defaultInstance; public static TableSchema getDefaultInstance() { @@ -126,20 +126,20 @@ public final class TableSchemaMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TableSchema( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -188,10 +188,10 @@ public final class TableSchemaMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { @@ -204,35 +204,35 @@ public final class TableSchemaMessage { makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.class, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public TableSchema parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new TableSchema(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } public interface AttributeOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string name = 1; /** @@ -246,7 +246,7 @@ public final class TableSchemaMessage { /** * required string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); // required string value = 2; @@ -261,21 +261,21 @@ public final class TableSchemaMessage { /** * required string value = 2; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute} */ public static final class Attribute extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements AttributeOrBuilder { // Use Attribute.newBuilder() to construct. - private Attribute(com.google.protobuf.GeneratedMessage.Builder builder) { + private Attribute(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Attribute(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Attribute(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Attribute defaultInstance; public static Attribute getDefaultInstance() { @@ -286,20 +286,20 @@ public final class TableSchemaMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Attribute( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -327,40 +327,40 @@ public final class TableSchemaMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.class, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Attribute parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Attribute(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -382,8 +382,8 @@ public final class TableSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -394,17 +394,17 @@ public final class TableSchemaMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -425,8 +425,8 @@ public final class TableSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { value_ = s; @@ -437,17 +437,17 @@ public final class TableSchemaMessage { /** * required string value = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -472,7 +472,7 @@ public final class TableSchemaMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -491,11 +491,11 @@ public final class TableSchemaMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, getValueBytes()); } size += getUnknownFields().getSerializedSize(); @@ -511,24 +511,24 @@ public final class TableSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom(java.io.InputStream input) @@ -537,7 +537,7 @@ public final class TableSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -547,18 +547,18 @@ public final class TableSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -572,7 +572,7 @@ public final class TableSchemaMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -580,14 +580,14 @@ public final class TableSchemaMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.AttributeOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -600,12 +600,12 @@ public final class TableSchemaMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -625,7 +625,7 @@ public final class TableSchemaMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor; } @@ -659,7 +659,7 @@ public final class TableSchemaMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute)other); } else { @@ -697,13 +697,13 @@ public final class TableSchemaMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute) e.getUnfinishedMessage(); throw e; } finally { @@ -729,7 +729,7 @@ public final class TableSchemaMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -740,17 +740,17 @@ public final class TableSchemaMessage { /** * required string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -779,7 +779,7 @@ public final class TableSchemaMessage { * required string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -803,7 +803,7 @@ public final class TableSchemaMessage { public java.lang.String getValue() { java.lang.Object ref = value_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); value_ = s; return s; @@ -814,17 +814,17 @@ public final class TableSchemaMessage { /** * required string value = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -853,7 +853,7 @@ public final class TableSchemaMessage { * required string value = 2; */ public Builder setValueBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -892,8 +892,8 @@ public final class TableSchemaMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -904,17 +904,17 @@ public final class TableSchemaMessage { /** * optional string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -1058,7 +1058,7 @@ public final class TableSchemaMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1086,23 +1086,23 @@ public final class TableSchemaMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } for (int i = 0; i < attrs_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(2, attrs_.get(i)); } for (int i = 0; i < columns_.size(); i++) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeMessageSize(3, columns_.get(i)); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBoolSize(4, inMemory_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBoolSize(5, readOnly_); } size += getUnknownFields().getSerializedSize(); @@ -1118,24 +1118,24 @@ public final class TableSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom(java.io.InputStream input) @@ -1144,7 +1144,7 @@ public final class TableSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1154,18 +1154,18 @@ public final class TableSchemaMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1179,7 +1179,7 @@ public final class TableSchemaMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1187,14 +1187,14 @@ public final class TableSchemaMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchemaOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1207,12 +1207,12 @@ public final class TableSchemaMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getAttrsFieldBuilder(); getColumnsFieldBuilder(); } @@ -1248,7 +1248,7 @@ public final class TableSchemaMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor; } @@ -1304,7 +1304,7 @@ public final class TableSchemaMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema)other); } else { @@ -1339,7 +1339,7 @@ public final class TableSchemaMessage { attrs_ = other.attrs_; bitField0_ = (bitField0_ & ~0x00000002); attrsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getAttrsFieldBuilder() : null; } else { attrsBuilder_.addAllMessages(other.attrs_); @@ -1365,7 +1365,7 @@ public final class TableSchemaMessage { columns_ = other.columns_; bitField0_ = (bitField0_ & ~0x00000004); columnsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getColumnsFieldBuilder() : null; } else { columnsBuilder_.addAllMessages(other.columns_); @@ -1399,13 +1399,13 @@ public final class TableSchemaMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema) e.getUnfinishedMessage(); throw e; } finally { @@ -1431,7 +1431,7 @@ public final class TableSchemaMessage { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -1442,17 +1442,17 @@ public final class TableSchemaMessage { /** * optional string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1481,7 +1481,7 @@ public final class TableSchemaMessage { * optional string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1501,7 +1501,7 @@ public final class TableSchemaMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.AttributeOrBuilder> attrsBuilder_; /** @@ -1716,11 +1716,11 @@ public final class TableSchemaMessage { getAttrsBuilderList() { return getAttrsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.AttributeOrBuilder> getAttrsFieldBuilder() { if (attrsBuilder_ == null) { - attrsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + attrsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.AttributeOrBuilder>( attrs_, ((bitField0_ & 0x00000002) == 0x00000002), @@ -1741,7 +1741,7 @@ public final class TableSchemaMessage { } } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchemaOrBuilder> columnsBuilder_; /** @@ -1956,11 +1956,11 @@ public final class TableSchemaMessage { getColumnsBuilderList() { return getColumnsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchemaOrBuilder> getColumnsFieldBuilder() { if (columnsBuilder_ == null) { - columnsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + columnsBuilder_ = new org.apache.hbase.shaded.com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchemaOrBuilder>( columns_, ((bitField0_ & 0x00000004) == 0x00000004), @@ -2064,22 +2064,22 @@ public final class TableSchemaMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -2094,29 +2094,29 @@ public final class TableSchemaMessage { "\010readOnly\030\005 \001(\010\032(\n\tAttribute\022\014\n\004name\030\001 \002" + "(\t\022\r\n\005value\030\002 \002(\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor, new java.lang.String[] { "Name", "Attrs", "Columns", "InMemory", "ReadOnly", }); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor, new java.lang.String[] { "Name", "Value", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.getDescriptor(), }, assigner); } diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java index 590b0d3..8f3391b 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.rest.protobuf.generated; public final class VersionMessage { private VersionMessage() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface VersionOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional string restVersion = 1; /** @@ -23,7 +23,7 @@ public final class VersionMessage { /** * optional string restVersion = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getRestVersionBytes(); // optional string jvmVersion = 2; @@ -38,7 +38,7 @@ public final class VersionMessage { /** * optional string jvmVersion = 2; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getJvmVersionBytes(); // optional string osVersion = 3; @@ -53,7 +53,7 @@ public final class VersionMessage { /** * optional string osVersion = 3; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getOsVersionBytes(); // optional string serverVersion = 4; @@ -68,7 +68,7 @@ public final class VersionMessage { /** * optional string serverVersion = 4; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getServerVersionBytes(); // optional string jerseyVersion = 5; @@ -83,21 +83,21 @@ public final class VersionMessage { /** * optional string jerseyVersion = 5; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getJerseyVersionBytes(); } /** * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Version} */ public static final class Version extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements VersionOrBuilder { // Use Version.newBuilder() to construct. - private Version(com.google.protobuf.GeneratedMessage.Builder builder) { + private Version(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private Version(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private Version(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Version defaultInstance; public static Version getDefaultInstance() { @@ -108,20 +108,20 @@ public final class VersionMessage { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Version( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -164,40 +164,40 @@ public final class VersionMessage { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.class, org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public Version parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new Version(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -219,8 +219,8 @@ public final class VersionMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { restVersion_ = s; @@ -231,17 +231,17 @@ public final class VersionMessage { /** * optional string restVersion = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRestVersionBytes() { java.lang.Object ref = restVersion_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); restVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -262,8 +262,8 @@ public final class VersionMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { jvmVersion_ = s; @@ -274,17 +274,17 @@ public final class VersionMessage { /** * optional string jvmVersion = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getJvmVersionBytes() { java.lang.Object ref = jvmVersion_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); jvmVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -305,8 +305,8 @@ public final class VersionMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { osVersion_ = s; @@ -317,17 +317,17 @@ public final class VersionMessage { /** * optional string osVersion = 3; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getOsVersionBytes() { java.lang.Object ref = osVersion_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); osVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -348,8 +348,8 @@ public final class VersionMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { serverVersion_ = s; @@ -360,17 +360,17 @@ public final class VersionMessage { /** * optional string serverVersion = 4; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getServerVersionBytes() { java.lang.Object ref = serverVersion_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); serverVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -391,8 +391,8 @@ public final class VersionMessage { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { jerseyVersion_ = s; @@ -403,17 +403,17 @@ public final class VersionMessage { /** * optional string jerseyVersion = 5; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getJerseyVersionBytes() { java.lang.Object ref = jerseyVersion_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); jerseyVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -433,7 +433,7 @@ public final class VersionMessage { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -461,23 +461,23 @@ public final class VersionMessage { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getRestVersionBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, getJvmVersionBytes()); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(3, getOsVersionBytes()); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(4, getServerVersionBytes()); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(5, getJerseyVersionBytes()); } size += getUnknownFields().getSerializedSize(); @@ -493,24 +493,24 @@ public final class VersionMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom(java.io.InputStream input) @@ -519,7 +519,7 @@ public final class VersionMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -529,18 +529,18 @@ public final class VersionMessage { } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -554,7 +554,7 @@ public final class VersionMessage { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -562,14 +562,14 @@ public final class VersionMessage { * Protobuf type {@code org.apache.hadoop.hbase.rest.protobuf.generated.Version} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.VersionOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -582,12 +582,12 @@ public final class VersionMessage { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -613,7 +613,7 @@ public final class VersionMessage { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor; } @@ -659,7 +659,7 @@ public final class VersionMessage { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version)other); } else { @@ -704,13 +704,13 @@ public final class VersionMessage { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version) e.getUnfinishedMessage(); throw e; } finally { @@ -736,7 +736,7 @@ public final class VersionMessage { public java.lang.String getRestVersion() { java.lang.Object ref = restVersion_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); restVersion_ = s; return s; @@ -747,17 +747,17 @@ public final class VersionMessage { /** * optional string restVersion = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRestVersionBytes() { java.lang.Object ref = restVersion_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); restVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -786,7 +786,7 @@ public final class VersionMessage { * optional string restVersion = 1; */ public Builder setRestVersionBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -810,7 +810,7 @@ public final class VersionMessage { public java.lang.String getJvmVersion() { java.lang.Object ref = jvmVersion_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); jvmVersion_ = s; return s; @@ -821,17 +821,17 @@ public final class VersionMessage { /** * optional string jvmVersion = 2; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getJvmVersionBytes() { java.lang.Object ref = jvmVersion_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); jvmVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -860,7 +860,7 @@ public final class VersionMessage { * optional string jvmVersion = 2; */ public Builder setJvmVersionBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -884,7 +884,7 @@ public final class VersionMessage { public java.lang.String getOsVersion() { java.lang.Object ref = osVersion_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); osVersion_ = s; return s; @@ -895,17 +895,17 @@ public final class VersionMessage { /** * optional string osVersion = 3; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getOsVersionBytes() { java.lang.Object ref = osVersion_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); osVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -934,7 +934,7 @@ public final class VersionMessage { * optional string osVersion = 3; */ public Builder setOsVersionBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -958,7 +958,7 @@ public final class VersionMessage { public java.lang.String getServerVersion() { java.lang.Object ref = serverVersion_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); serverVersion_ = s; return s; @@ -969,17 +969,17 @@ public final class VersionMessage { /** * optional string serverVersion = 4; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getServerVersionBytes() { java.lang.Object ref = serverVersion_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); serverVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1008,7 +1008,7 @@ public final class VersionMessage { * optional string serverVersion = 4; */ public Builder setServerVersionBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1032,7 +1032,7 @@ public final class VersionMessage { public java.lang.String getJerseyVersion() { java.lang.Object ref = jerseyVersion_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); jerseyVersion_ = s; return s; @@ -1043,17 +1043,17 @@ public final class VersionMessage { /** * optional string jerseyVersion = 5; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getJerseyVersionBytes() { java.lang.Object ref = jerseyVersion_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); jerseyVersion_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1082,7 +1082,7 @@ public final class VersionMessage { * optional string jerseyVersion = 5; */ public Builder setJerseyVersionBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1103,17 +1103,17 @@ public final class VersionMessage { // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.Version) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1123,23 +1123,23 @@ public final class VersionMessage { "\001(\t\022\021\n\tosVersion\030\003 \001(\t\022\025\n\rserverVersion\030" + "\004 \001(\t\022\025\n\rjerseyVersion\030\005 \001(\t" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor, new java.lang.String[] { "RestVersion", "JvmVersion", "OsVersion", "ServerVersion", "JerseyVersion", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminClient.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminClient.java index a7f14f7..7c3350d 100644 --- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminClient.java +++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminClient.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hbase.rsgroup; import com.google.common.collect.Sets; import com.google.common.net.HostAndPort; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import java.io.IOException; import java.util.ArrayList; diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java index 22bad72..6628abe 100644 --- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java +++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java @@ -22,9 +22,9 @@ package org.apache.hadoop.hbase.rsgroup; import com.google.common.collect.Sets; import com.google.common.net.HostAndPort; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; import java.io.IOException; import java.util.HashSet; diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java index 01efefc..c97033f 100644 --- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java +++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java @@ -25,7 +25,7 @@ import com.google.common.collect.Maps; import com.google.common.collect.Sets; import com.google.common.net.HostAndPort; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import java.io.IOException; import java.util.ArrayList; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/HTableWrapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/HTableWrapper.java index 7865cc0..9097698 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/HTableWrapper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/HTableWrapper.java @@ -37,10 +37,10 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; import org.apache.hadoop.io.MultipleIOException; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * A wrapper for HTable. Can be used to restrict privilege. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java index 5d3cbc2..58637f1 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.ProcessRequest; import org.apache.hadoop.hbase.regionserver.RowProcessor; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * Convenience class that is used to make RowProcessorEndpoint invocations. * For example usage, refer TestRowProcessorEndpoint diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java index cc78626..9c53374 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java @@ -39,11 +39,11 @@ import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRespo import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateService; import org.apache.hadoop.hbase.regionserver.InternalScanner; -import com.google.protobuf.ByteString; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * A concrete AggregateProtocol implementation. Its system level coprocessor diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java index ab5fc78..b413123 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java @@ -34,11 +34,11 @@ import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcesso import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.RowProcessor; -import com.google.protobuf.ByteString; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * This class demonstrates how to implement atomic read-modify-writes diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java index caf6a14..19f1d27 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorService.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hbase.coprocessor; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Service; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.HBaseInterfaceAudience; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/EndpointObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/EndpointObserver.java index 1076437..68bcf94 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/EndpointObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/EndpointObserver.java @@ -26,8 +26,8 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.Coprocessor; import org.apache.hadoop.hbase.HBaseInterfaceAudience; -import com.google.protobuf.Message; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Coprocessors implement this interface to observe and mediate endpoint invocations diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java index e771a92..3750a8a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java @@ -41,9 +41,9 @@ import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MutateR import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MutateRowsResponse; import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MultiRowMutationService; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * This class demonstrates how to implement atomic multi row transactions using diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/SingletonCoprocessorService.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/SingletonCoprocessorService.java index 88db6b6..98cd444 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/SingletonCoprocessorService.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/SingletonCoprocessorService.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hbase.coprocessor; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Service; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.HBaseInterfaceAudience; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java index 2224414..3429784 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.ForeignExc import org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage; import org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.StackTraceElementMessage; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A ForeignException is an exception from another thread or process. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java index a9cf0f1..e609973 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java @@ -33,7 +33,7 @@ import org.apache.hadoop.util.StringUtils; import org.apache.htrace.Trace; import org.apache.htrace.TraceScope; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * The request processing logic, which is usually executed in thread pools provided by an diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/PriorityFunction.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/PriorityFunction.java index f56bf6f..42a12e7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/PriorityFunction.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/PriorityFunction.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.ipc; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.HBaseInterfaceAudience; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RWQueueRpcExecutor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RWQueueRpcExecutor.java index e0203ab..f48fad1 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RWQueueRpcExecutor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RWQueueRpcExecutor.java @@ -40,7 +40,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanRequest; import org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RequestHeader; import org.apache.hadoop.hbase.util.ReflectionUtils; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * RPC Executor that uses different queues for reads and writes. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java index f0aed2e..34ba767 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java @@ -131,13 +131,13 @@ import org.apache.htrace.TraceInfo; import org.codehaus.jackson.map.ObjectMapper; import com.google.common.util.concurrent.ThreadFactoryBuilder; -import com.google.protobuf.BlockingService; -import com.google.protobuf.CodedInputStream; -import com.google.protobuf.CodedOutputStream; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.ServiceException; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; /** * An RPC server that hosts protobuf described Services. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java index ab8b485..f900e62 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServerInterface.java @@ -31,10 +31,10 @@ import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.security.authorize.PolicyProvider; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.BlockingService; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @InterfaceAudience.LimitedPrivate({HBaseInterfaceAudience.COPROC, HBaseInterfaceAudience.PHOENIX}) @InterfaceStability.Evolving diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java index 37e4e44..f17bcbd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java @@ -18,7 +18,7 @@ */ package org.apache.hadoop.hbase.mapreduce; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import com.codahale.metrics.MetricRegistry; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -798,7 +798,7 @@ public class TableMapReduceUtil { // pull necessary dependencies org.apache.zookeeper.ZooKeeper.class, io.netty.channel.Channel.class, - com.google.protobuf.Message.class, + org.apache.hbase.shaded.com.google.protobuf.Message.class, com.google.common.collect.Lists.class, org.apache.htrace.Trace.class, com.codahale.metrics.MetricRegistry.class); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ExpiredMobFileCleanerChore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ExpiredMobFileCleanerChore.java index a21edcc..eda50ef 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ExpiredMobFileCleanerChore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ExpiredMobFileCleanerChore.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.mob.ExpiredMobFileCleaner; import org.apache.hadoop.hbase.mob.MobConstants; import org.apache.hadoop.hbase.mob.MobUtils; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * The Class ExpiredMobFileCleanerChore for running cleaner regularly to remove the expired diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java index 3b5af42..f5f6bba 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java @@ -21,8 +21,8 @@ package org.apache.hadoop.hbase.master; import com.google.common.annotations.VisibleForTesting; import com.google.common.collect.Lists; import com.google.common.collect.Maps; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Service; import java.io.IOException; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterAnnotationReadingPriorityFunction.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterAnnotationReadingPriorityFunction.java index 1e6dade..a3bdd81 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterAnnotationReadingPriorityFunction.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterAnnotationReadingPriorityFunction.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.master; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java index 6ee022f..1332c9e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java @@ -95,13 +95,13 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Pair; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.ByteString; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Implements the master RPC services. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java index d6802fe..46f6db1 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java @@ -40,7 +40,7 @@ import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; import org.apache.hadoop.hbase.quotas.MasterQuotaManager; import org.apache.hadoop.hbase.security.User; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Services Master supplies @@ -263,7 +263,7 @@ public interface MasterServices extends Server { * *

* Only a single instance may be registered for a given {@link Service} subclass (the - * instances are keyed on {@link com.google.protobuf.Descriptors.ServiceDescriptor#getFullName()}. + * instances are keyed on {@link org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor#getFullName()}. * After the first registration, subsequent calls with the same service name will fail with * a return value of {@code false}. *

diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java index dabef71..ad64fc8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java @@ -78,8 +78,8 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ByteString; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * The ServerManager class manages info about region servers. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java index c884806..70e5faf 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java @@ -49,7 +49,7 @@ import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Helper class for {@link FavoredNodeLoadBalancer} that has all the intelligence diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java index 13b5fab..5d0b363 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java @@ -18,7 +18,7 @@ */ package org.apache.hadoop.hbase.master.normalizer; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/ExpiredMobFileCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/ExpiredMobFileCleaner.java index 3c965cb..ae4c4ee 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/ExpiredMobFileCleaner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/ExpiredMobFileCleaner.java @@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * The cleaner to delete the expired MOB files. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/Sweeper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/Sweeper.java index 8547c8c..0b32a5f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/Sweeper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/Sweeper.java @@ -38,7 +38,7 @@ import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * The sweep tool. It deletes the mob files that are not used and merges the small mob files to diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java index 5961645..a91cdf5 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.monitoring; import org.apache.hadoop.hbase.classification.InterfaceAudience; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * A MonitoredTask implementation optimized for use with RPC Handlers diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java index a29595b..f0c0a1ed 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.Operation; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * A MonitoredTask implementation designed for use with RPC Handlers diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinatorRpcs.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinatorRpcs.java index 085d642..fb05e79 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinatorRpcs.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinatorRpcs.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.zookeeper.ZKUtil; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * ZooKeeper based {@link ProcedureCoordinatorRpcs} for a {@link ProcedureCoordinator} diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureMemberRpcs.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureMemberRpcs.java index 2e03a60..f480ea3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureMemberRpcs.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureMemberRpcs.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.zookeeper.ZKUtil; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * ZooKeeper based controller for a procedure member. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java index 8cb2237..f7afde8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java @@ -46,7 +46,7 @@ import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.wal.WAL.Entry; import org.apache.hadoop.hbase.wal.WALKey; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @InterfaceAudience.Private public class ReplicationProtbufUtil { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.java index 1eacc75..777519a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.java @@ -41,8 +41,8 @@ import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier; import org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RequestHeader; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.Message; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; import org.apache.hadoop.hbase.security.User; /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java index 65375b8..28f1a02 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java @@ -28,7 +28,7 @@ import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * Base class for RowProcessor with some default implementations. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 5524de0..8eccbc7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -189,13 +189,13 @@ import com.google.common.base.Preconditions; import com.google.common.collect.Lists; import com.google.common.collect.Maps; import com.google.common.io.Closeables; -import com.google.protobuf.ByteString; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; @SuppressWarnings("deprecation") @InterfaceAudience.Private diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java index 1476190..2d8485a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java @@ -185,13 +185,13 @@ import org.apache.zookeeper.data.Stat; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; import com.google.common.collect.Maps; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.Descriptors; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import sun.misc.Signal; import sun.misc.SignalHandler; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java index 2d27219..7f1fe38 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java @@ -186,11 +186,11 @@ import org.apache.hadoop.hbase.zookeeper.ZKSplitLog; import org.apache.zookeeper.KeeperException; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ByteString; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; /** * Implements the regionserver RPC services. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Region.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Region.java index 9b1f82a..e2c4797 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Region.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Region.java @@ -18,9 +18,9 @@ package org.apache.hadoop.hbase.regionserver; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; import java.io.IOException; import java.util.Collection; import java.util.List; @@ -551,7 +551,7 @@ public interface Region extends ConfigurationObserver { /** * Executes a single protocol buffer coprocessor endpoint {@link Service} method using * the registered protocol handlers. {@link Service} implementations must be registered via the - * {@link Region#registerService(com.google.protobuf.Service)} + * {@link Region#registerService(org.apache.hbase.shaded.com.google.protobuf.Service)} * method before they are available. * * @param controller an {@code RpcContoller} implementation to pass to the invoked service @@ -560,19 +560,19 @@ public interface Region extends ConfigurationObserver { * @return a protocol buffer {@code Message} instance containing the method's result * @throws IOException if no registered service handler is found or an error * occurs during the invocation - * @see org.apache.hadoop.hbase.regionserver.Region#registerService(com.google.protobuf.Service) + * @see org.apache.hadoop.hbase.regionserver.Region#registerService(org.apache.hbase.shaded.com.google.protobuf.Service) */ Message execService(RpcController controller, CoprocessorServiceCall call) throws IOException; /** * Registers a new protocol buffer {@link Service} subclass as a coprocessor endpoint to * be available for handling - * {@link Region#execService(com.google.protobuf.RpcController, + * {@link Region#execService(org.apache.hbase.shaded.com.google.protobuf.RpcController, * org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall)}} calls. * *

* Only a single instance may be registered per region for a given {@link Service} subclass (the - * instances are keyed on {@link com.google.protobuf.Descriptors.ServiceDescriptor#getFullName()}. + * instances are keyed on {@link org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor#getFullName()}. * After the first registration, subsequent calls with the same service name will fail with * a return value of {@code false}. *

diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 48e67f7..18ce325 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -31,8 +31,8 @@ import java.util.regex.Matcher; import com.google.common.collect.ImmutableList; import com.google.common.collect.Lists; -import com.google.protobuf.Message; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; import org.apache.commons.collections.map.AbstractReferenceMap; import org.apache.commons.collections.map.ReferenceMap; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java index c6689a9..06d79a4 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java @@ -38,7 +38,7 @@ import org.apache.hadoop.hbase.regionserver.throttle.ThroughputController; import org.apache.hadoop.hbase.wal.WAL; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Services provided by {@link HRegionServer} diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RowProcessor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RowProcessor.java index 34901b7..428a285 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RowProcessor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RowProcessor.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; /** * Defines the procedure to atomically perform multiple scans and mutations diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java index 537329a..a1d0675 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java @@ -57,7 +57,7 @@ import org.apache.hadoop.hbase.regionserver.RegionServerServices; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * This manager class handles the work dealing with snapshots for a {@link HRegionServer}. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogReader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogReader.java index 0755358..5d51873 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogReader.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogReader.java @@ -44,8 +44,8 @@ import org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALTrailer; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.wal.WAL.Entry; -import com.google.protobuf.CodedInputStream; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * A Protobuf based WAL has the following structure: diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.java index 6b89e89..40c5101 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.java @@ -40,7 +40,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.ReflectionUtils; import org.apache.hadoop.io.IOUtils; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEditsReplaySink.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEditsReplaySink.java index 9d5e052..9e90390 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEditsReplaySink.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEditsReplaySink.java @@ -50,7 +50,7 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.wal.WAL.Entry; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * This class is responsible for replaying the edits coming from a failed region server. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALUtil.java index 197144d..9100afc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALUtil.java @@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl; import org.apache.hadoop.hbase.wal.WAL; import org.apache.hadoop.hbase.wal.WALKey; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; /** * Helper methods to ease Region Server integration with the Write Ahead Log (WAL). diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/RegionReplicaReplicationEndpoint.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/RegionReplicaReplicationEndpoint.java index da37cfa..7adfee6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/RegionReplicaReplicationEndpoint.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/RegionReplicaReplicationEndpoint.java @@ -79,7 +79,7 @@ import org.apache.hadoop.util.StringUtils; import com.google.common.cache.Cache; import com.google.common.cache.CacheBuilder; import com.google.common.collect.Lists; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * A {@link org.apache.hadoop.hbase.replication.ReplicationEndpoint} endpoint diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java index fb19a96..43354a7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java @@ -128,10 +128,10 @@ import com.google.common.collect.Lists; import com.google.common.collect.MapMaker; import com.google.common.collect.Maps; import com.google.common.collect.Sets; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Provides basic authorization checks for data access and administrative diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/HbaseObjectWritableFor96Migration.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/HbaseObjectWritableFor96Migration.java index d0d9b63..0ff5212 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/HbaseObjectWritableFor96Migration.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/HbaseObjectWritableFor96Migration.java @@ -99,8 +99,8 @@ import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.WritableFactories; import org.apache.hadoop.io.WritableUtils; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; /** *

This is a customized version of the polymorphic hadoop diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/SecureBulkLoadEndpoint.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/SecureBulkLoadEndpoint.java index c1f9251..d43c96f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/SecureBulkLoadEndpoint.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/SecureBulkLoadEndpoint.java @@ -18,9 +18,9 @@ package org.apache.hadoop.hbase.security.access; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java index a7e6113..25a5d67 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java @@ -20,9 +20,9 @@ package org.apache.hadoop.hbase.security.token; import java.io.IOException; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.classification.InterfaceAudience; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java index 68817bc..b697344 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java @@ -22,7 +22,7 @@ import java.io.IOException; import java.lang.reflect.UndeclaredThrowableException; import java.security.PrivilegedExceptionAction; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.classification.InterfaceAudience; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java index 6e2f8ed..db9f891 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java @@ -109,10 +109,10 @@ import org.apache.hadoop.hbase.util.Pair; import com.google.common.collect.Lists; import com.google.common.collect.MapMaker; -import com.google.protobuf.ByteString; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Coprocessor that has both the MasterObserver and RegionObserver implemented that supports in diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java index df0c348..fe405f3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java @@ -18,8 +18,8 @@ package org.apache.hadoop.hbase.snapshot; -import com.google.protobuf.CodedInputStream; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import java.io.FileNotFoundException; import java.io.IOException; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java index 3bb3575..93a4c48 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hbase.snapshot; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import java.io.IOException; import java.io.InterruptedIOException; import java.util.ArrayList; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputHelper.java index 2225191..d94233b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputHelper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputHelper.java @@ -59,7 +59,7 @@ import java.util.concurrent.TimeUnit; import com.google.common.base.Throwables; import com.google.common.collect.ImmutableMap; -import com.google.protobuf.CodedOutputStream; +import org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputSaslHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputSaslHelper.java index 341d4ec..4d1735c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputSaslHelper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FanOutOneBlockAsyncDFSOutputSaslHelper.java @@ -18,24 +18,6 @@ package org.apache.hadoop.hbase.util; import static io.netty.handler.timeout.IdleState.READER_IDLE; -import io.netty.buffer.ByteBuf; -import io.netty.buffer.ByteBufOutputStream; -import io.netty.buffer.CompositeByteBuf; -import io.netty.buffer.Unpooled; -import io.netty.channel.Channel; -import io.netty.channel.ChannelDuplexHandler; -import io.netty.channel.ChannelHandlerContext; -import io.netty.channel.ChannelOutboundHandlerAdapter; -import io.netty.channel.ChannelPipeline; -import io.netty.channel.ChannelPromise; -import io.netty.channel.SimpleChannelInboundHandler; -import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -import io.netty.handler.codec.MessageToByteEncoder; -import io.netty.handler.codec.protobuf.ProtobufDecoder; -import io.netty.handler.codec.protobuf.ProtobufVarint32FrameDecoder; -import io.netty.handler.timeout.IdleStateEvent; -import io.netty.handler.timeout.IdleStateHandler; -import io.netty.util.concurrent.Promise; import java.io.IOException; import java.lang.reflect.Constructor; @@ -63,13 +45,6 @@ import javax.security.sasl.Sasl; import javax.security.sasl.SaslClient; import javax.security.sasl.SaslException; -import com.google.common.base.Charsets; -import com.google.common.collect.ImmutableSet; -import com.google.common.collect.Lists; -import com.google.common.collect.Maps; -import com.google.protobuf.ByteString; -import com.google.protobuf.CodedOutputStream; - import org.apache.commons.codec.binary.Base64; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -89,6 +64,31 @@ import org.apache.hadoop.security.SaslPropertiesResolver; import org.apache.hadoop.security.SaslRpcServer.QualityOfProtection; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.token.Token; +import org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream; + +import com.google.common.base.Charsets; +import com.google.common.collect.ImmutableSet; +import com.google.common.collect.Lists; +import com.google.common.collect.Maps; + +import io.netty.buffer.ByteBuf; +import io.netty.buffer.ByteBufOutputStream; +import io.netty.buffer.CompositeByteBuf; +import io.netty.buffer.Unpooled; +import io.netty.channel.Channel; +import io.netty.channel.ChannelDuplexHandler; +import io.netty.channel.ChannelHandlerContext; +import io.netty.channel.ChannelOutboundHandlerAdapter; +import io.netty.channel.ChannelPipeline; +import io.netty.channel.ChannelPromise; +import io.netty.channel.SimpleChannelInboundHandler; +import io.netty.handler.codec.LengthFieldBasedFrameDecoder; +import io.netty.handler.codec.MessageToByteEncoder; +import io.netty.handler.codec.protobuf.ProtobufDecoder; +import io.netty.handler.codec.protobuf.ProtobufVarint32FrameDecoder; +import io.netty.handler.timeout.IdleStateEvent; +import io.netty.handler.timeout.IdleStateHandler; +import io.netty.util.concurrent.Promise; /** * Helper class for adding sasl support for {@link FanOutOneBlockAsyncDFSOutput}. @@ -654,7 +654,10 @@ public final class FanOutOneBlockAsyncDFSOutputSaslHelper { DataTransferEncryptorMessageProto.newBuilder(); builder.setStatus(DataTransferEncryptorStatus.SUCCESS); if (payload != null) { - builder.setPayload(ByteString.copyFrom(payload)); + // NOTE!!! Explicit reference to HDFS's transitively included protobuf! HBase has + // shaded and relocated the protobuf it uses to be at + // org.apache.hbase.shaded.com.google.protobuf. + builder.setPayload(com.google.protobuf.ByteString.copyFrom(payload)); } if (options != null) { CIPHER_HELPER.addCipherOptions(builder, options); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java index 9abef9c..03de0e0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java @@ -61,7 +61,7 @@ import com.google.common.collect.Lists; import com.google.common.collect.Multimap; import com.google.common.collect.Ordering; import com.google.common.collect.TreeMultimap; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.io.IOUtils; import org.apache.commons.lang.RandomStringUtils; import org.apache.commons.lang.StringUtils; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java index 3c2203b..549d45a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java @@ -15,7 +15,7 @@ * See the License for the specific language governing permissions and * limitations under the License. * - * The portion of this file denoted by 'Copied from com.google.protobuf.CodedInputStream' + * The portion of this file denoted by 'Copied from org.apache.hbase.shaded.com.google.protobuf.CodedInputStream' * is from Protocol Buffers v2.4.1 under the following license * * Copyright 2008 Google Inc. All rights reserved. @@ -66,7 +66,7 @@ public abstract class ProtoUtil { * @throws IOException if it is malformed or EOF. */ public static int readRawVarint32(DataInput in) throws IOException { - // Copied from com.google.protobuf.CodedInputStream v2.4.1 readRawVarint32 + // Copied from org.apache.hbase.shaded.com.google.protobuf.CodedInputStream v2.4.1 readRawVarint32 byte tmp = in.readByte(); if (tmp >= 0) { return tmp; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALKey.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALKey.java index 86fdfbd..ec9b11d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALKey.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALKey.java @@ -50,7 +50,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * A Key for an entry in the WAL. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java index 8d78480..289d6ad 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java @@ -122,8 +122,8 @@ import org.apache.hadoop.ipc.RemoteException; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; import com.google.common.collect.Lists; -import com.google.protobuf.ServiceException; -import com.google.protobuf.TextFormat; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.TextFormat; /** * This class is responsible for splitting up a bunch of regionserver commit log diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java index 6cd1963..bfc936d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java @@ -54,7 +54,7 @@ import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Basic mock region server services. Should only be instantiated by HBaseTestingUtility.b diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/QosTestHelper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/QosTestHelper.java index 6db201f..0cf0047 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/QosTestHelper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/QosTestHelper.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.protobuf.generated.RPCProtos; import org.apache.hadoop.hbase.regionserver.AnnotationReadingPriorityFunction; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java index 227db6f..61c72f7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java @@ -48,8 +48,8 @@ import org.mockito.Mockito; import org.mockito.invocation.InvocationOnMock; import org.mockito.stubbing.Answer; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test MetaTableAccessor but without spinning up a cluster. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableLocator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableLocator.java index ba6e1d4..a6b6883 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableLocator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableLocator.java @@ -54,8 +54,8 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.mockito.Mockito; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test {@link org.apache.hadoop.hbase.zookeeper.MetaTableLocator} diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java index e6c17a5..ff9270f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({MiscTests.class, SmallTests.class}) public class TestServerLoad { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java index 10dbed0..41ae8f9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java @@ -70,7 +70,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Class to test HBaseAdmin. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java index e510d28..855497b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java @@ -66,7 +66,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java index 515e763..9835e5d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java @@ -45,8 +45,8 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test the scenario where a HRegionServer#scan() call, while scanning, timeout at client side and diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java index 36276fa..62a3efc 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java @@ -46,11 +46,11 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({MediumTests.class, ClientTests.class}) public class TestClientTimeouts { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java index 45093bb..edabdb3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java @@ -58,8 +58,8 @@ import org.mockito.Mockito; import org.mockito.invocation.InvocationOnMock; import org.mockito.stubbing.Answer; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({SmallTests.class, ClientTests.class}) public class TestHBaseAdminNoCluster { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaCache.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaCache.java index 23b9eed..d8237b5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaCache.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaCache.java @@ -17,8 +17,8 @@ */ package org.apache.hadoop.hbase.client; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.*; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java index 4e4ff5e..9e6f718 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java @@ -36,9 +36,9 @@ import org.apache.hadoop.hbase.protobuf.ResponseConverter; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java index 6c0ea49..7e8a287 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java @@ -37,9 +37,9 @@ import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Test coprocessor endpoint that always returns {@code null} for requests to the last region diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java index 32d978c..b380b7e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java @@ -38,9 +38,9 @@ import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.util.Bytes; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; /** * Test coprocessor endpoint that always throws a {@link DoNotRetryIOException} for requests on diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ProtobufCoprocessorService.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ProtobufCoprocessorService.java index cdda28a..1a09d80 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ProtobufCoprocessorService.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/ProtobufCoprocessorService.java @@ -18,9 +18,9 @@ package org.apache.hadoop.hbase.coprocessor; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; import org.apache.hadoop.hbase.Coprocessor; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java index d62e950..41a6961 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java @@ -50,7 +50,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * TestEndpoint: test cases to verify the batch execution of coprocessor Endpoint diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java index 1768a2a..987d871 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java @@ -60,8 +60,8 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * TestEndpoint: test cases to verify coprocessor Endpoint diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java index 7695361..a56993d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java @@ -42,7 +42,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({CoprocessorTests.class, MediumTests.class}) public class TestCoprocessorTableEndpoint { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java index 1484c34..89b8127 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java @@ -40,9 +40,9 @@ import org.junit.AfterClass; import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertTrue; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java index b3d3890..c9db9b9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java @@ -74,7 +74,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java index 61b47ff..6c36ac1 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.coprocessor.protobuf.generated; public final class ColumnAggregationProtos { private ColumnAggregationProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface SumRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes family = 1; /** @@ -19,7 +19,7 @@ public final class ColumnAggregationProtos { /** * required bytes family = 1; */ - com.google.protobuf.ByteString getFamily(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily(); // optional bytes qualifier = 2; /** @@ -29,20 +29,20 @@ public final class ColumnAggregationProtos { /** * optional bytes qualifier = 2; */ - com.google.protobuf.ByteString getQualifier(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier(); } /** * Protobuf type {@code SumRequest} */ public static final class SumRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements SumRequestOrBuilder { // Use SumRequest.newBuilder() to construct. - private SumRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private SumRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SumRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SumRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SumRequest defaultInstance; public static SumRequest getDefaultInstance() { @@ -53,20 +53,20 @@ public final class ColumnAggregationProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SumRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -94,47 +94,47 @@ public final class ColumnAggregationProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public SumRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new SumRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes family = 1; public static final int FAMILY_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString family_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString family_; /** * required bytes family = 1; */ @@ -144,13 +144,13 @@ public final class ColumnAggregationProtos { /** * required bytes family = 1; */ - public com.google.protobuf.ByteString getFamily() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily() { return family_; } // optional bytes qualifier = 2; public static final int QUALIFIER_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString qualifier_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString qualifier_; /** * optional bytes qualifier = 2; */ @@ -160,13 +160,13 @@ public final class ColumnAggregationProtos { /** * optional bytes qualifier = 2; */ - public com.google.protobuf.ByteString getQualifier() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier() { return qualifier_; } private void initFields() { - family_ = com.google.protobuf.ByteString.EMPTY; - qualifier_ = com.google.protobuf.ByteString.EMPTY; + family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -181,7 +181,7 @@ public final class ColumnAggregationProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -200,11 +200,11 @@ public final class ColumnAggregationProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, family_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, qualifier_); } size += getUnknownFields().getSerializedSize(); @@ -267,24 +267,24 @@ public final class ColumnAggregationProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(java.io.InputStream input) @@ -293,7 +293,7 @@ public final class ColumnAggregationProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -303,18 +303,18 @@ public final class ColumnAggregationProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -328,7 +328,7 @@ public final class ColumnAggregationProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -336,14 +336,14 @@ public final class ColumnAggregationProtos { * Protobuf type {@code SumRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -356,12 +356,12 @@ public final class ColumnAggregationProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -370,9 +370,9 @@ public final class ColumnAggregationProtos { public Builder clear() { super.clear(); - family_ = com.google.protobuf.ByteString.EMPTY; + family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - qualifier_ = com.google.protobuf.ByteString.EMPTY; + qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } @@ -381,7 +381,7 @@ public final class ColumnAggregationProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_descriptor; } @@ -415,7 +415,7 @@ public final class ColumnAggregationProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest)other); } else { @@ -445,13 +445,13 @@ public final class ColumnAggregationProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -464,7 +464,7 @@ public final class ColumnAggregationProtos { private int bitField0_; // required bytes family = 1; - private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes family = 1; */ @@ -474,13 +474,13 @@ public final class ColumnAggregationProtos { /** * required bytes family = 1; */ - public com.google.protobuf.ByteString getFamily() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily() { return family_; } /** * required bytes family = 1; */ - public Builder setFamily(com.google.protobuf.ByteString value) { + public Builder setFamily(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -500,7 +500,7 @@ public final class ColumnAggregationProtos { } // optional bytes qualifier = 2; - private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes qualifier = 2; */ @@ -510,13 +510,13 @@ public final class ColumnAggregationProtos { /** * optional bytes qualifier = 2; */ - public com.google.protobuf.ByteString getQualifier() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier() { return qualifier_; } /** * optional bytes qualifier = 2; */ - public Builder setQualifier(com.google.protobuf.ByteString value) { + public Builder setQualifier(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -547,7 +547,7 @@ public final class ColumnAggregationProtos { } public interface SumResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int64 sum = 1; /** @@ -563,14 +563,14 @@ public final class ColumnAggregationProtos { * Protobuf type {@code SumResponse} */ public static final class SumResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements SumResponseOrBuilder { // Use SumResponse.newBuilder() to construct. - private SumResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private SumResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SumResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SumResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SumResponse defaultInstance; public static SumResponse getDefaultInstance() { @@ -581,20 +581,20 @@ public final class ColumnAggregationProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SumResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -617,40 +617,40 @@ public final class ColumnAggregationProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public SumResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new SumResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -687,7 +687,7 @@ public final class ColumnAggregationProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -703,7 +703,7 @@ public final class ColumnAggregationProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(1, sum_); } size += getUnknownFields().getSerializedSize(); @@ -757,24 +757,24 @@ public final class ColumnAggregationProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(java.io.InputStream input) @@ -783,7 +783,7 @@ public final class ColumnAggregationProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -793,18 +793,18 @@ public final class ColumnAggregationProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -818,7 +818,7 @@ public final class ColumnAggregationProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -826,14 +826,14 @@ public final class ColumnAggregationProtos { * Protobuf type {@code SumResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -846,12 +846,12 @@ public final class ColumnAggregationProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -869,7 +869,7 @@ public final class ColumnAggregationProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_descriptor; } @@ -899,7 +899,7 @@ public final class ColumnAggregationProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse)other); } else { @@ -926,13 +926,13 @@ public final class ColumnAggregationProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -992,7 +992,7 @@ public final class ColumnAggregationProtos { * Protobuf service {@code ColumnAggregationService} */ public static abstract class ColumnAggregationService - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected ColumnAggregationService() {} public interface Interface { @@ -1000,39 +1000,39 @@ public final class ColumnAggregationProtos { * rpc sum(.SumRequest) returns (.SumResponse); */ public abstract void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new ColumnAggregationService() { @java.lang.Override public void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.sum(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -1046,9 +1046,9 @@ public final class ColumnAggregationProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1062,9 +1062,9 @@ public final class ColumnAggregationProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1085,26 +1085,26 @@ public final class ColumnAggregationProtos { * rpc sum(.SumRequest) returns (.SumResponse); */ public abstract void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -1113,7 +1113,7 @@ public final class ColumnAggregationProtos { switch(method.getIndex()) { case 0: this.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -1121,9 +1121,9 @@ public final class ColumnAggregationProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1137,9 +1137,9 @@ public final class ColumnAggregationProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1154,31 +1154,31 @@ public final class ColumnAggregationProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.ColumnAggregationService implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance())); @@ -1186,28 +1186,28 @@ public final class ColumnAggregationProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -1220,22 +1220,22 @@ public final class ColumnAggregationProtos { // @@protoc_insertion_point(class_scope:ColumnAggregationService) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_SumRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SumRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_SumResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SumResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1247,29 +1247,29 @@ public final class ColumnAggregationProtos { "coprocessor.protobuf.generatedB\027ColumnAg" + "gregationProtos\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_SumRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_SumRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SumRequest_descriptor, new java.lang.String[] { "Family", "Qualifier", }); internal_static_SumResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_SumResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SumResponse_descriptor, new java.lang.String[] { "Sum", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java index 6768aa2..a17774e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.coprocessor.protobuf.generated; public final class ColumnAggregationWithErrorsProtos { private ColumnAggregationWithErrorsProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface SumRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes family = 1; /** @@ -19,7 +19,7 @@ public final class ColumnAggregationWithErrorsProtos { /** * required bytes family = 1; */ - com.google.protobuf.ByteString getFamily(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily(); // optional bytes qualifier = 2; /** @@ -29,20 +29,20 @@ public final class ColumnAggregationWithErrorsProtos { /** * optional bytes qualifier = 2; */ - com.google.protobuf.ByteString getQualifier(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier(); } /** * Protobuf type {@code SumRequest} */ public static final class SumRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements SumRequestOrBuilder { // Use SumRequest.newBuilder() to construct. - private SumRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private SumRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SumRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SumRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SumRequest defaultInstance; public static SumRequest getDefaultInstance() { @@ -53,20 +53,20 @@ public final class ColumnAggregationWithErrorsProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SumRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -94,47 +94,47 @@ public final class ColumnAggregationWithErrorsProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public SumRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new SumRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes family = 1; public static final int FAMILY_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString family_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString family_; /** * required bytes family = 1; */ @@ -144,13 +144,13 @@ public final class ColumnAggregationWithErrorsProtos { /** * required bytes family = 1; */ - public com.google.protobuf.ByteString getFamily() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily() { return family_; } // optional bytes qualifier = 2; public static final int QUALIFIER_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString qualifier_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString qualifier_; /** * optional bytes qualifier = 2; */ @@ -160,13 +160,13 @@ public final class ColumnAggregationWithErrorsProtos { /** * optional bytes qualifier = 2; */ - public com.google.protobuf.ByteString getQualifier() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier() { return qualifier_; } private void initFields() { - family_ = com.google.protobuf.ByteString.EMPTY; - qualifier_ = com.google.protobuf.ByteString.EMPTY; + family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -181,7 +181,7 @@ public final class ColumnAggregationWithErrorsProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -200,11 +200,11 @@ public final class ColumnAggregationWithErrorsProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, family_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, qualifier_); } size += getUnknownFields().getSerializedSize(); @@ -267,24 +267,24 @@ public final class ColumnAggregationWithErrorsProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom(java.io.InputStream input) @@ -293,7 +293,7 @@ public final class ColumnAggregationWithErrorsProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -303,18 +303,18 @@ public final class ColumnAggregationWithErrorsProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -328,7 +328,7 @@ public final class ColumnAggregationWithErrorsProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -336,14 +336,14 @@ public final class ColumnAggregationWithErrorsProtos { * Protobuf type {@code SumRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -356,12 +356,12 @@ public final class ColumnAggregationWithErrorsProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -370,9 +370,9 @@ public final class ColumnAggregationWithErrorsProtos { public Builder clear() { super.clear(); - family_ = com.google.protobuf.ByteString.EMPTY; + family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - qualifier_ = com.google.protobuf.ByteString.EMPTY; + qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } @@ -381,7 +381,7 @@ public final class ColumnAggregationWithErrorsProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumRequest_descriptor; } @@ -415,7 +415,7 @@ public final class ColumnAggregationWithErrorsProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest)other); } else { @@ -445,13 +445,13 @@ public final class ColumnAggregationWithErrorsProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -464,7 +464,7 @@ public final class ColumnAggregationWithErrorsProtos { private int bitField0_; // required bytes family = 1; - private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes family = 1; */ @@ -474,13 +474,13 @@ public final class ColumnAggregationWithErrorsProtos { /** * required bytes family = 1; */ - public com.google.protobuf.ByteString getFamily() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily() { return family_; } /** * required bytes family = 1; */ - public Builder setFamily(com.google.protobuf.ByteString value) { + public Builder setFamily(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -500,7 +500,7 @@ public final class ColumnAggregationWithErrorsProtos { } // optional bytes qualifier = 2; - private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes qualifier = 2; */ @@ -510,13 +510,13 @@ public final class ColumnAggregationWithErrorsProtos { /** * optional bytes qualifier = 2; */ - public com.google.protobuf.ByteString getQualifier() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier() { return qualifier_; } /** * optional bytes qualifier = 2; */ - public Builder setQualifier(com.google.protobuf.ByteString value) { + public Builder setQualifier(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -547,7 +547,7 @@ public final class ColumnAggregationWithErrorsProtos { } public interface SumResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int64 sum = 1; /** @@ -563,14 +563,14 @@ public final class ColumnAggregationWithErrorsProtos { * Protobuf type {@code SumResponse} */ public static final class SumResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements SumResponseOrBuilder { // Use SumResponse.newBuilder() to construct. - private SumResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private SumResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SumResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SumResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SumResponse defaultInstance; public static SumResponse getDefaultInstance() { @@ -581,20 +581,20 @@ public final class ColumnAggregationWithErrorsProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SumResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -617,40 +617,40 @@ public final class ColumnAggregationWithErrorsProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public SumResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new SumResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -687,7 +687,7 @@ public final class ColumnAggregationWithErrorsProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -703,7 +703,7 @@ public final class ColumnAggregationWithErrorsProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(1, sum_); } size += getUnknownFields().getSerializedSize(); @@ -757,24 +757,24 @@ public final class ColumnAggregationWithErrorsProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom(java.io.InputStream input) @@ -783,7 +783,7 @@ public final class ColumnAggregationWithErrorsProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -793,18 +793,18 @@ public final class ColumnAggregationWithErrorsProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -818,7 +818,7 @@ public final class ColumnAggregationWithErrorsProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -826,14 +826,14 @@ public final class ColumnAggregationWithErrorsProtos { * Protobuf type {@code SumResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -846,12 +846,12 @@ public final class ColumnAggregationWithErrorsProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -869,7 +869,7 @@ public final class ColumnAggregationWithErrorsProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_SumResponse_descriptor; } @@ -899,7 +899,7 @@ public final class ColumnAggregationWithErrorsProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse)other); } else { @@ -926,13 +926,13 @@ public final class ColumnAggregationWithErrorsProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -992,7 +992,7 @@ public final class ColumnAggregationWithErrorsProtos { * Protobuf service {@code ColumnAggregationServiceWithErrors} */ public static abstract class ColumnAggregationServiceWithErrors - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected ColumnAggregationServiceWithErrors() {} public interface Interface { @@ -1000,39 +1000,39 @@ public final class ColumnAggregationWithErrorsProtos { * rpc sum(.SumRequest) returns (.SumResponse); */ public abstract void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new ColumnAggregationServiceWithErrors() { @java.lang.Override public void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.sum(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -1046,9 +1046,9 @@ public final class ColumnAggregationWithErrorsProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1062,9 +1062,9 @@ public final class ColumnAggregationWithErrorsProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1085,26 +1085,26 @@ public final class ColumnAggregationWithErrorsProtos { * rpc sum(.SumRequest) returns (.SumResponse); */ public abstract void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -1113,7 +1113,7 @@ public final class ColumnAggregationWithErrorsProtos { switch(method.getIndex()) { case 0: this.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -1121,9 +1121,9 @@ public final class ColumnAggregationWithErrorsProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1137,9 +1137,9 @@ public final class ColumnAggregationWithErrorsProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1154,31 +1154,31 @@ public final class ColumnAggregationWithErrorsProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationServiceWithErrors implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse.getDefaultInstance())); @@ -1186,28 +1186,28 @@ public final class ColumnAggregationWithErrorsProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.SumResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -1220,22 +1220,22 @@ public final class ColumnAggregationWithErrorsProtos { // @@protoc_insertion_point(class_scope:ColumnAggregationServiceWithErrors) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_SumRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SumRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_SumResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SumResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1248,29 +1248,29 @@ public final class ColumnAggregationWithErrorsProtos { ".generatedB!ColumnAggregationWithErrorsP" + "rotos\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_SumRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_SumRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SumRequest_descriptor, new java.lang.String[] { "Family", "Qualifier", }); internal_static_SumResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_SumResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SumResponse_descriptor, new java.lang.String[] { "Sum", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java index c0b1917..78d6f39 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.coprocessor.protobuf.generated; public final class ColumnAggregationWithNullResponseProtos { private ColumnAggregationWithNullResponseProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface SumRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes family = 1; /** @@ -19,7 +19,7 @@ public final class ColumnAggregationWithNullResponseProtos { /** * required bytes family = 1; */ - com.google.protobuf.ByteString getFamily(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily(); // optional bytes qualifier = 2; /** @@ -29,20 +29,20 @@ public final class ColumnAggregationWithNullResponseProtos { /** * optional bytes qualifier = 2; */ - com.google.protobuf.ByteString getQualifier(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier(); } /** * Protobuf type {@code SumRequest} */ public static final class SumRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements SumRequestOrBuilder { // Use SumRequest.newBuilder() to construct. - private SumRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private SumRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SumRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SumRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SumRequest defaultInstance; public static SumRequest getDefaultInstance() { @@ -53,20 +53,20 @@ public final class ColumnAggregationWithNullResponseProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SumRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -94,47 +94,47 @@ public final class ColumnAggregationWithNullResponseProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public SumRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new SumRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes family = 1; public static final int FAMILY_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString family_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString family_; /** * required bytes family = 1; */ @@ -144,13 +144,13 @@ public final class ColumnAggregationWithNullResponseProtos { /** * required bytes family = 1; */ - public com.google.protobuf.ByteString getFamily() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily() { return family_; } // optional bytes qualifier = 2; public static final int QUALIFIER_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString qualifier_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString qualifier_; /** * optional bytes qualifier = 2; */ @@ -160,13 +160,13 @@ public final class ColumnAggregationWithNullResponseProtos { /** * optional bytes qualifier = 2; */ - public com.google.protobuf.ByteString getQualifier() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier() { return qualifier_; } private void initFields() { - family_ = com.google.protobuf.ByteString.EMPTY; - qualifier_ = com.google.protobuf.ByteString.EMPTY; + family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -181,7 +181,7 @@ public final class ColumnAggregationWithNullResponseProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -200,11 +200,11 @@ public final class ColumnAggregationWithNullResponseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, family_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, qualifier_); } size += getUnknownFields().getSerializedSize(); @@ -267,24 +267,24 @@ public final class ColumnAggregationWithNullResponseProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom(java.io.InputStream input) @@ -293,7 +293,7 @@ public final class ColumnAggregationWithNullResponseProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -303,18 +303,18 @@ public final class ColumnAggregationWithNullResponseProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -328,7 +328,7 @@ public final class ColumnAggregationWithNullResponseProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -336,14 +336,14 @@ public final class ColumnAggregationWithNullResponseProtos { * Protobuf type {@code SumRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -356,12 +356,12 @@ public final class ColumnAggregationWithNullResponseProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -370,9 +370,9 @@ public final class ColumnAggregationWithNullResponseProtos { public Builder clear() { super.clear(); - family_ = com.google.protobuf.ByteString.EMPTY; + family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - qualifier_ = com.google.protobuf.ByteString.EMPTY; + qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } @@ -381,7 +381,7 @@ public final class ColumnAggregationWithNullResponseProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumRequest_descriptor; } @@ -415,7 +415,7 @@ public final class ColumnAggregationWithNullResponseProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest)other); } else { @@ -445,13 +445,13 @@ public final class ColumnAggregationWithNullResponseProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -464,7 +464,7 @@ public final class ColumnAggregationWithNullResponseProtos { private int bitField0_; // required bytes family = 1; - private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString family_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes family = 1; */ @@ -474,13 +474,13 @@ public final class ColumnAggregationWithNullResponseProtos { /** * required bytes family = 1; */ - public com.google.protobuf.ByteString getFamily() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getFamily() { return family_; } /** * required bytes family = 1; */ - public Builder setFamily(com.google.protobuf.ByteString value) { + public Builder setFamily(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -500,7 +500,7 @@ public final class ColumnAggregationWithNullResponseProtos { } // optional bytes qualifier = 2; - private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString qualifier_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * optional bytes qualifier = 2; */ @@ -510,13 +510,13 @@ public final class ColumnAggregationWithNullResponseProtos { /** * optional bytes qualifier = 2; */ - public com.google.protobuf.ByteString getQualifier() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getQualifier() { return qualifier_; } /** * optional bytes qualifier = 2; */ - public Builder setQualifier(com.google.protobuf.ByteString value) { + public Builder setQualifier(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -547,7 +547,7 @@ public final class ColumnAggregationWithNullResponseProtos { } public interface SumResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional int64 sum = 1; /** @@ -563,14 +563,14 @@ public final class ColumnAggregationWithNullResponseProtos { * Protobuf type {@code SumResponse} */ public static final class SumResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements SumResponseOrBuilder { // Use SumResponse.newBuilder() to construct. - private SumResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private SumResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SumResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SumResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SumResponse defaultInstance; public static SumResponse getDefaultInstance() { @@ -581,20 +581,20 @@ public final class ColumnAggregationWithNullResponseProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SumResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -617,40 +617,40 @@ public final class ColumnAggregationWithNullResponseProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public SumResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new SumResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -683,7 +683,7 @@ public final class ColumnAggregationWithNullResponseProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -699,7 +699,7 @@ public final class ColumnAggregationWithNullResponseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt64Size(1, sum_); } size += getUnknownFields().getSerializedSize(); @@ -753,24 +753,24 @@ public final class ColumnAggregationWithNullResponseProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom(java.io.InputStream input) @@ -779,7 +779,7 @@ public final class ColumnAggregationWithNullResponseProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -789,18 +789,18 @@ public final class ColumnAggregationWithNullResponseProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -814,7 +814,7 @@ public final class ColumnAggregationWithNullResponseProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -822,14 +822,14 @@ public final class ColumnAggregationWithNullResponseProtos { * Protobuf type {@code SumResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -842,12 +842,12 @@ public final class ColumnAggregationWithNullResponseProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -865,7 +865,7 @@ public final class ColumnAggregationWithNullResponseProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_SumResponse_descriptor; } @@ -895,7 +895,7 @@ public final class ColumnAggregationWithNullResponseProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse)other); } else { @@ -918,13 +918,13 @@ public final class ColumnAggregationWithNullResponseProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -984,7 +984,7 @@ public final class ColumnAggregationWithNullResponseProtos { * Protobuf service {@code ColumnAggregationServiceNullResponse} */ public static abstract class ColumnAggregationServiceNullResponse - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected ColumnAggregationServiceNullResponse() {} public interface Interface { @@ -992,39 +992,39 @@ public final class ColumnAggregationWithNullResponseProtos { * rpc sum(.SumRequest) returns (.SumResponse); */ public abstract void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new ColumnAggregationServiceNullResponse() { @java.lang.Override public void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.sum(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -1038,9 +1038,9 @@ public final class ColumnAggregationWithNullResponseProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1054,9 +1054,9 @@ public final class ColumnAggregationWithNullResponseProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1077,26 +1077,26 @@ public final class ColumnAggregationWithNullResponseProtos { * rpc sum(.SumRequest) returns (.SumResponse); */ public abstract void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -1105,7 +1105,7 @@ public final class ColumnAggregationWithNullResponseProtos { switch(method.getIndex()) { case 0: this.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -1113,9 +1113,9 @@ public final class ColumnAggregationWithNullResponseProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1129,9 +1129,9 @@ public final class ColumnAggregationWithNullResponseProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1146,31 +1146,31 @@ public final class ColumnAggregationWithNullResponseProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationServiceNullResponse implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse.getDefaultInstance())); @@ -1178,28 +1178,28 @@ public final class ColumnAggregationWithNullResponseProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse sum( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.SumResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -1212,22 +1212,22 @@ public final class ColumnAggregationWithNullResponseProtos { // @@protoc_insertion_point(class_scope:ColumnAggregationServiceNullResponse) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_SumRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SumRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_SumResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SumResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1240,29 +1240,29 @@ public final class ColumnAggregationWithNullResponseProtos { "obuf.generatedB\'ColumnAggregationWithNul" + "lResponseProtos\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_SumRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_SumRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SumRequest_descriptor, new java.lang.String[] { "Family", "Qualifier", }); internal_static_SumResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_SumResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SumResponse_descriptor, new java.lang.String[] { "Sum", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java index 12ce551..56254e6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java @@ -6,23 +6,23 @@ package org.apache.hadoop.hbase.coprocessor.protobuf.generated; public final class DummyRegionServerEndpointProtos { private DummyRegionServerEndpointProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface DummyRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code DummyRequest} */ public static final class DummyRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements DummyRequestOrBuilder { // Use DummyRequest.newBuilder() to construct. - private DummyRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private DummyRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private DummyRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private DummyRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final DummyRequest defaultInstance; public static DummyRequest getDefaultInstance() { @@ -33,19 +33,19 @@ public final class DummyRegionServerEndpointProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DummyRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -63,40 +63,40 @@ public final class DummyRegionServerEndpointProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public DummyRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new DummyRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -111,7 +111,7 @@ public final class DummyRegionServerEndpointProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -165,24 +165,24 @@ public final class DummyRegionServerEndpointProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(java.io.InputStream input) @@ -191,7 +191,7 @@ public final class DummyRegionServerEndpointProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -201,18 +201,18 @@ public final class DummyRegionServerEndpointProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -226,7 +226,7 @@ public final class DummyRegionServerEndpointProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -234,14 +234,14 @@ public final class DummyRegionServerEndpointProtos { * Protobuf type {@code DummyRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -254,12 +254,12 @@ public final class DummyRegionServerEndpointProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -275,7 +275,7 @@ public final class DummyRegionServerEndpointProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyRequest_descriptor; } @@ -298,7 +298,7 @@ public final class DummyRegionServerEndpointProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)other); } else { @@ -318,13 +318,13 @@ public final class DummyRegionServerEndpointProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -347,7 +347,7 @@ public final class DummyRegionServerEndpointProtos { } public interface DummyResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string value = 1; /** @@ -361,21 +361,21 @@ public final class DummyRegionServerEndpointProtos { /** * required string value = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes(); } /** * Protobuf type {@code DummyResponse} */ public static final class DummyResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements DummyResponseOrBuilder { // Use DummyResponse.newBuilder() to construct. - private DummyResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private DummyResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private DummyResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private DummyResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final DummyResponse defaultInstance; public static DummyResponse getDefaultInstance() { @@ -386,20 +386,20 @@ public final class DummyRegionServerEndpointProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DummyResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -422,40 +422,40 @@ public final class DummyRegionServerEndpointProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public DummyResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new DummyResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -477,8 +477,8 @@ public final class DummyRegionServerEndpointProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { value_ = s; @@ -489,17 +489,17 @@ public final class DummyRegionServerEndpointProtos { /** * required string value = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -519,7 +519,7 @@ public final class DummyRegionServerEndpointProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -535,7 +535,7 @@ public final class DummyRegionServerEndpointProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getValueBytes()); } size += getUnknownFields().getSerializedSize(); @@ -589,24 +589,24 @@ public final class DummyRegionServerEndpointProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(java.io.InputStream input) @@ -615,7 +615,7 @@ public final class DummyRegionServerEndpointProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -625,18 +625,18 @@ public final class DummyRegionServerEndpointProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -650,7 +650,7 @@ public final class DummyRegionServerEndpointProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -658,14 +658,14 @@ public final class DummyRegionServerEndpointProtos { * Protobuf type {@code DummyResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -678,12 +678,12 @@ public final class DummyRegionServerEndpointProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -701,7 +701,7 @@ public final class DummyRegionServerEndpointProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_DummyResponse_descriptor; } @@ -731,7 +731,7 @@ public final class DummyRegionServerEndpointProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse)other); } else { @@ -760,13 +760,13 @@ public final class DummyRegionServerEndpointProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -792,7 +792,7 @@ public final class DummyRegionServerEndpointProtos { public java.lang.String getValue() { java.lang.Object ref = value_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); value_ = s; return s; @@ -803,17 +803,17 @@ public final class DummyRegionServerEndpointProtos { /** * required string value = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); value_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -842,7 +842,7 @@ public final class DummyRegionServerEndpointProtos { * required string value = 1; */ public Builder setValueBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -867,7 +867,7 @@ public final class DummyRegionServerEndpointProtos { * Protobuf service {@code DummyService} */ public static abstract class DummyService - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected DummyService() {} public interface Interface { @@ -875,55 +875,55 @@ public final class DummyRegionServerEndpointProtos { * rpc dummyCall(.DummyRequest) returns (.DummyResponse); */ public abstract void dummyCall( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc dummyThrow(.DummyRequest) returns (.DummyResponse); */ public abstract void dummyThrow( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new DummyService() { @java.lang.Override public void dummyCall( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.dummyCall(controller, request, done); } @java.lang.Override public void dummyThrow( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.dummyThrow(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -939,9 +939,9 @@ public final class DummyRegionServerEndpointProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -957,9 +957,9 @@ public final class DummyRegionServerEndpointProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -982,34 +982,34 @@ public final class DummyRegionServerEndpointProtos { * rpc dummyCall(.DummyRequest) returns (.DummyResponse); */ public abstract void dummyCall( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc dummyThrow(.DummyRequest) returns (.DummyResponse); */ public abstract void dummyThrow( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -1018,12 +1018,12 @@ public final class DummyRegionServerEndpointProtos { switch(method.getIndex()) { case 0: this.dummyCall(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.dummyThrow(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -1031,9 +1031,9 @@ public final class DummyRegionServerEndpointProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -1049,9 +1049,9 @@ public final class DummyRegionServerEndpointProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -1068,46 +1068,46 @@ public final class DummyRegionServerEndpointProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyService implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void dummyCall( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance())); } public void dummyThrow( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance())); @@ -1115,33 +1115,33 @@ public final class DummyRegionServerEndpointProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyCall( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyThrow( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyCall( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -1151,9 +1151,9 @@ public final class DummyRegionServerEndpointProtos { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyThrow( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, @@ -1166,22 +1166,22 @@ public final class DummyRegionServerEndpointProtos { // @@protoc_insertion_point(class_scope:DummyService) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_DummyRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_DummyRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_DummyResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_DummyResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1194,29 +1194,29 @@ public final class DummyRegionServerEndpointProtos { "atedB\037DummyRegionServerEndpointProtos\210\001\001" + "\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_DummyRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_DummyRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_DummyRequest_descriptor, new java.lang.String[] { }); internal_static_DummyResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_DummyResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_DummyResponse_descriptor, new java.lang.String[] { "Value", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java index 7ba5b8e..56ab5bd 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.coprocessor.protobuf.generated; public final class IncrementCounterProcessorTestProtos { private IncrementCounterProcessorTestProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface IncCounterProcessorRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes row = 1; /** @@ -19,7 +19,7 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 1; */ - com.google.protobuf.ByteString getRow(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getRow(); // required int32 counter = 2; /** @@ -35,14 +35,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code IncCounterProcessorRequest} */ public static final class IncCounterProcessorRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements IncCounterProcessorRequestOrBuilder { // Use IncCounterProcessorRequest.newBuilder() to construct. - private IncCounterProcessorRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private IncCounterProcessorRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private IncCounterProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private IncCounterProcessorRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final IncCounterProcessorRequest defaultInstance; public static IncCounterProcessorRequest getDefaultInstance() { @@ -53,20 +53,20 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IncCounterProcessorRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -94,47 +94,47 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public IncCounterProcessorRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new IncCounterProcessorRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes row = 1; public static final int ROW_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString row_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_; /** * required bytes row = 1; */ @@ -144,7 +144,7 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 1; */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } @@ -165,7 +165,7 @@ public final class IncrementCounterProcessorTestProtos { } private void initFields() { - row_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; counter_ = 0; } private byte memoizedIsInitialized = -1; @@ -185,7 +185,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -204,11 +204,11 @@ public final class IncrementCounterProcessorTestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, row_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(2, counter_); } size += getUnknownFields().getSerializedSize(); @@ -271,24 +271,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(java.io.InputStream input) @@ -297,7 +297,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -307,18 +307,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -332,7 +332,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -340,14 +340,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code IncCounterProcessorRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -360,12 +360,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -374,7 +374,7 @@ public final class IncrementCounterProcessorTestProtos { public Builder clear() { super.clear(); - row_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); counter_ = 0; bitField0_ = (bitField0_ & ~0x00000002); @@ -385,7 +385,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor; } @@ -419,7 +419,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest)other); } else { @@ -453,13 +453,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -472,7 +472,7 @@ public final class IncrementCounterProcessorTestProtos { private int bitField0_; // required bytes row = 1; - private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes row = 1; */ @@ -482,13 +482,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 1; */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } /** * required bytes row = 1; */ - public Builder setRow(com.google.protobuf.ByteString value) { + public Builder setRow(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -552,7 +552,7 @@ public final class IncrementCounterProcessorTestProtos { } public interface IncCounterProcessorResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int32 response = 1; /** @@ -568,14 +568,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code IncCounterProcessorResponse} */ public static final class IncCounterProcessorResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements IncCounterProcessorResponseOrBuilder { // Use IncCounterProcessorResponse.newBuilder() to construct. - private IncCounterProcessorResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private IncCounterProcessorResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private IncCounterProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private IncCounterProcessorResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final IncCounterProcessorResponse defaultInstance; public static IncCounterProcessorResponse getDefaultInstance() { @@ -586,20 +586,20 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IncCounterProcessorResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -622,40 +622,40 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public IncCounterProcessorResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new IncCounterProcessorResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -692,7 +692,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -708,7 +708,7 @@ public final class IncrementCounterProcessorTestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(1, response_); } size += getUnknownFields().getSerializedSize(); @@ -762,24 +762,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(java.io.InputStream input) @@ -788,7 +788,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -798,18 +798,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -823,7 +823,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -831,14 +831,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code IncCounterProcessorResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -851,12 +851,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -874,7 +874,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor; } @@ -904,7 +904,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse)other); } else { @@ -931,13 +931,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -994,7 +994,7 @@ public final class IncrementCounterProcessorTestProtos { } public interface FriendsOfFriendsProcessorRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes person = 1; /** @@ -1004,7 +1004,7 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes person = 1; */ - com.google.protobuf.ByteString getPerson(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getPerson(); // required bytes row = 2; /** @@ -1014,7 +1014,7 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 2; */ - com.google.protobuf.ByteString getRow(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getRow(); // repeated string result = 3; /** @@ -1033,21 +1033,21 @@ public final class IncrementCounterProcessorTestProtos { /** * repeated string result = 3; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getResultBytes(int index); } /** * Protobuf type {@code FriendsOfFriendsProcessorRequest} */ public static final class FriendsOfFriendsProcessorRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements FriendsOfFriendsProcessorRequestOrBuilder { // Use FriendsOfFriendsProcessorRequest.newBuilder() to construct. - private FriendsOfFriendsProcessorRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private FriendsOfFriendsProcessorRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private FriendsOfFriendsProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private FriendsOfFriendsProcessorRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final FriendsOfFriendsProcessorRequest defaultInstance; public static FriendsOfFriendsProcessorRequest getDefaultInstance() { @@ -1058,20 +1058,20 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private FriendsOfFriendsProcessorRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1099,7 +1099,7 @@ public final class IncrementCounterProcessorTestProtos { } case 26: { if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - result_ = new com.google.protobuf.LazyStringArrayList(); + result_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000004; } result_.add(input.readBytes()); @@ -1107,50 +1107,50 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - result_ = new com.google.protobuf.UnmodifiableLazyStringList(result_); + result_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList(result_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public FriendsOfFriendsProcessorRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new FriendsOfFriendsProcessorRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes person = 1; public static final int PERSON_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString person_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString person_; /** * required bytes person = 1; */ @@ -1160,13 +1160,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes person = 1; */ - public com.google.protobuf.ByteString getPerson() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getPerson() { return person_; } // required bytes row = 2; public static final int ROW_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString row_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_; /** * required bytes row = 2; */ @@ -1176,13 +1176,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 2; */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } // repeated string result = 3; public static final int RESULT_FIELD_NUMBER = 3; - private com.google.protobuf.LazyStringList result_; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList result_; /** * repeated string result = 3; */ @@ -1205,15 +1205,15 @@ public final class IncrementCounterProcessorTestProtos { /** * repeated string result = 3; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getResultBytes(int index) { return result_.getByteString(index); } private void initFields() { - person_ = com.google.protobuf.ByteString.EMPTY; - row_ = com.google.protobuf.ByteString.EMPTY; - result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + person_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -1232,7 +1232,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1254,17 +1254,17 @@ public final class IncrementCounterProcessorTestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, person_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, row_); } { int dataSize = 0; for (int i = 0; i < result_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(result_.getByteString(i)); } size += dataSize; @@ -1336,24 +1336,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(java.io.InputStream input) @@ -1362,7 +1362,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1372,18 +1372,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1397,7 +1397,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1405,14 +1405,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code FriendsOfFriendsProcessorRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1425,12 +1425,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1439,11 +1439,11 @@ public final class IncrementCounterProcessorTestProtos { public Builder clear() { super.clear(); - person_ = com.google.protobuf.ByteString.EMPTY; + person_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - row_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); - result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000004); return this; } @@ -1452,7 +1452,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor; } @@ -1482,7 +1482,7 @@ public final class IncrementCounterProcessorTestProtos { } result.row_ = row_; if (((bitField0_ & 0x00000004) == 0x00000004)) { - result_ = new com.google.protobuf.UnmodifiableLazyStringList( + result_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList( result_); bitField0_ = (bitField0_ & ~0x00000004); } @@ -1492,7 +1492,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest)other); } else { @@ -1536,13 +1536,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -1555,7 +1555,7 @@ public final class IncrementCounterProcessorTestProtos { private int bitField0_; // required bytes person = 1; - private com.google.protobuf.ByteString person_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString person_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes person = 1; */ @@ -1565,13 +1565,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes person = 1; */ - public com.google.protobuf.ByteString getPerson() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getPerson() { return person_; } /** * required bytes person = 1; */ - public Builder setPerson(com.google.protobuf.ByteString value) { + public Builder setPerson(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1591,7 +1591,7 @@ public final class IncrementCounterProcessorTestProtos { } // required bytes row = 2; - private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes row = 2; */ @@ -1601,13 +1601,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 2; */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } /** * required bytes row = 2; */ - public Builder setRow(com.google.protobuf.ByteString value) { + public Builder setRow(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1627,10 +1627,10 @@ public final class IncrementCounterProcessorTestProtos { } // repeated string result = 3; - private com.google.protobuf.LazyStringList result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureResultIsMutable() { if (!((bitField0_ & 0x00000004) == 0x00000004)) { - result_ = new com.google.protobuf.LazyStringArrayList(result_); + result_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(result_); bitField0_ |= 0x00000004; } } @@ -1656,7 +1656,7 @@ public final class IncrementCounterProcessorTestProtos { /** * repeated string result = 3; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getResultBytes(int index) { return result_.getByteString(index); } @@ -1700,7 +1700,7 @@ public final class IncrementCounterProcessorTestProtos { * repeated string result = 3; */ public Builder clearResult() { - result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000004); onChanged(); return this; @@ -1709,7 +1709,7 @@ public final class IncrementCounterProcessorTestProtos { * repeated string result = 3; */ public Builder addResultBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1731,7 +1731,7 @@ public final class IncrementCounterProcessorTestProtos { } public interface FriendsOfFriendsProcessorResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // repeated string result = 1; /** @@ -1750,21 +1750,21 @@ public final class IncrementCounterProcessorTestProtos { /** * repeated string result = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getResultBytes(int index); } /** * Protobuf type {@code FriendsOfFriendsProcessorResponse} */ public static final class FriendsOfFriendsProcessorResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements FriendsOfFriendsProcessorResponseOrBuilder { // Use FriendsOfFriendsProcessorResponse.newBuilder() to construct. - private FriendsOfFriendsProcessorResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private FriendsOfFriendsProcessorResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private FriendsOfFriendsProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private FriendsOfFriendsProcessorResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final FriendsOfFriendsProcessorResponse defaultInstance; public static FriendsOfFriendsProcessorResponse getDefaultInstance() { @@ -1775,20 +1775,20 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private FriendsOfFriendsProcessorResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1806,7 +1806,7 @@ public final class IncrementCounterProcessorTestProtos { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - result_ = new com.google.protobuf.LazyStringArrayList(); + result_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000001; } result_.add(input.readBytes()); @@ -1814,49 +1814,49 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - result_ = new com.google.protobuf.UnmodifiableLazyStringList(result_); + result_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList(result_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public FriendsOfFriendsProcessorResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new FriendsOfFriendsProcessorResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } // repeated string result = 1; public static final int RESULT_FIELD_NUMBER = 1; - private com.google.protobuf.LazyStringList result_; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList result_; /** * repeated string result = 1; */ @@ -1879,13 +1879,13 @@ public final class IncrementCounterProcessorTestProtos { /** * repeated string result = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getResultBytes(int index) { return result_.getByteString(index); } private void initFields() { - result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -1896,7 +1896,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (int i = 0; i < result_.size(); i++) { @@ -1914,7 +1914,7 @@ public final class IncrementCounterProcessorTestProtos { { int dataSize = 0; for (int i = 0; i < result_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream + dataSize += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(result_.getByteString(i)); } size += dataSize; @@ -1968,24 +1968,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(java.io.InputStream input) @@ -1994,7 +1994,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2004,18 +2004,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2029,7 +2029,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2037,14 +2037,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code FriendsOfFriendsProcessorResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2057,12 +2057,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -2071,7 +2071,7 @@ public final class IncrementCounterProcessorTestProtos { public Builder clear() { super.clear(); - result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); return this; } @@ -2080,7 +2080,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor; } @@ -2101,7 +2101,7 @@ public final class IncrementCounterProcessorTestProtos { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) == 0x00000001)) { - result_ = new com.google.protobuf.UnmodifiableLazyStringList( + result_ = new org.apache.hbase.shaded.com.google.protobuf.UnmodifiableLazyStringList( result_); bitField0_ = (bitField0_ & ~0x00000001); } @@ -2110,7 +2110,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse)other); } else { @@ -2140,13 +2140,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -2159,10 +2159,10 @@ public final class IncrementCounterProcessorTestProtos { private int bitField0_; // repeated string result = 1; - private com.google.protobuf.LazyStringList result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.LazyStringList result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureResultIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - result_ = new com.google.protobuf.LazyStringArrayList(result_); + result_ = new org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList(result_); bitField0_ |= 0x00000001; } } @@ -2188,7 +2188,7 @@ public final class IncrementCounterProcessorTestProtos { /** * repeated string result = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getResultBytes(int index) { return result_.getByteString(index); } @@ -2232,7 +2232,7 @@ public final class IncrementCounterProcessorTestProtos { * repeated string result = 1; */ public Builder clearResult() { - result_ = com.google.protobuf.LazyStringArrayList.EMPTY; + result_ = org.apache.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; @@ -2241,7 +2241,7 @@ public final class IncrementCounterProcessorTestProtos { * repeated string result = 1; */ public Builder addResultBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -2263,7 +2263,7 @@ public final class IncrementCounterProcessorTestProtos { } public interface RowSwapProcessorRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes row1 = 1; /** @@ -2273,7 +2273,7 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row1 = 1; */ - com.google.protobuf.ByteString getRow1(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getRow1(); // required bytes row2 = 2; /** @@ -2283,20 +2283,20 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row2 = 2; */ - com.google.protobuf.ByteString getRow2(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getRow2(); } /** * Protobuf type {@code RowSwapProcessorRequest} */ public static final class RowSwapProcessorRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements RowSwapProcessorRequestOrBuilder { // Use RowSwapProcessorRequest.newBuilder() to construct. - private RowSwapProcessorRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private RowSwapProcessorRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private RowSwapProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private RowSwapProcessorRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final RowSwapProcessorRequest defaultInstance; public static RowSwapProcessorRequest getDefaultInstance() { @@ -2307,20 +2307,20 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RowSwapProcessorRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -2348,47 +2348,47 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public RowSwapProcessorRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new RowSwapProcessorRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes row1 = 1; public static final int ROW1_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString row1_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row1_; /** * required bytes row1 = 1; */ @@ -2398,13 +2398,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row1 = 1; */ - public com.google.protobuf.ByteString getRow1() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow1() { return row1_; } // required bytes row2 = 2; public static final int ROW2_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString row2_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row2_; /** * required bytes row2 = 2; */ @@ -2414,13 +2414,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row2 = 2; */ - public com.google.protobuf.ByteString getRow2() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow2() { return row2_; } private void initFields() { - row1_ = com.google.protobuf.ByteString.EMPTY; - row2_ = com.google.protobuf.ByteString.EMPTY; + row1_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; + row2_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -2439,7 +2439,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -2458,11 +2458,11 @@ public final class IncrementCounterProcessorTestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, row1_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(2, row2_); } size += getUnknownFields().getSerializedSize(); @@ -2525,24 +2525,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom(java.io.InputStream input) @@ -2551,7 +2551,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2561,18 +2561,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2586,7 +2586,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2594,14 +2594,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code RowSwapProcessorRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2614,12 +2614,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -2628,9 +2628,9 @@ public final class IncrementCounterProcessorTestProtos { public Builder clear() { super.clear(); - row1_ = com.google.protobuf.ByteString.EMPTY; + row1_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - row2_ = com.google.protobuf.ByteString.EMPTY; + row2_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } @@ -2639,7 +2639,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorRequest_descriptor; } @@ -2673,7 +2673,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest)other); } else { @@ -2707,13 +2707,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -2726,7 +2726,7 @@ public final class IncrementCounterProcessorTestProtos { private int bitField0_; // required bytes row1 = 1; - private com.google.protobuf.ByteString row1_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row1_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes row1 = 1; */ @@ -2736,13 +2736,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row1 = 1; */ - public com.google.protobuf.ByteString getRow1() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow1() { return row1_; } /** * required bytes row1 = 1; */ - public Builder setRow1(com.google.protobuf.ByteString value) { + public Builder setRow1(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -2762,7 +2762,7 @@ public final class IncrementCounterProcessorTestProtos { } // required bytes row2 = 2; - private com.google.protobuf.ByteString row2_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row2_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes row2 = 2; */ @@ -2772,13 +2772,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row2 = 2; */ - public com.google.protobuf.ByteString getRow2() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow2() { return row2_; } /** * required bytes row2 = 2; */ - public Builder setRow2(com.google.protobuf.ByteString value) { + public Builder setRow2(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -2809,20 +2809,20 @@ public final class IncrementCounterProcessorTestProtos { } public interface RowSwapProcessorResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code RowSwapProcessorResponse} */ public static final class RowSwapProcessorResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements RowSwapProcessorResponseOrBuilder { // Use RowSwapProcessorResponse.newBuilder() to construct. - private RowSwapProcessorResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private RowSwapProcessorResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private RowSwapProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private RowSwapProcessorResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final RowSwapProcessorResponse defaultInstance; public static RowSwapProcessorResponse getDefaultInstance() { @@ -2833,19 +2833,19 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RowSwapProcessorResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -2863,40 +2863,40 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public RowSwapProcessorResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new RowSwapProcessorResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -2911,7 +2911,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -2965,24 +2965,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom(java.io.InputStream input) @@ -2991,7 +2991,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3001,18 +3001,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3026,7 +3026,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -3034,14 +3034,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code RowSwapProcessorResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -3054,12 +3054,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -3075,7 +3075,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_RowSwapProcessorResponse_descriptor; } @@ -3098,7 +3098,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse)other); } else { @@ -3118,13 +3118,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -3147,7 +3147,7 @@ public final class IncrementCounterProcessorTestProtos { } public interface TimeoutProcessorRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required bytes row = 1; /** @@ -3157,20 +3157,20 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 1; */ - com.google.protobuf.ByteString getRow(); + org.apache.hbase.shaded.com.google.protobuf.ByteString getRow(); } /** * Protobuf type {@code TimeoutProcessorRequest} */ public static final class TimeoutProcessorRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements TimeoutProcessorRequestOrBuilder { // Use TimeoutProcessorRequest.newBuilder() to construct. - private TimeoutProcessorRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private TimeoutProcessorRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private TimeoutProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private TimeoutProcessorRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TimeoutProcessorRequest defaultInstance; public static TimeoutProcessorRequest getDefaultInstance() { @@ -3181,20 +3181,20 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TimeoutProcessorRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -3217,47 +3217,47 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public TimeoutProcessorRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new TimeoutProcessorRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; // required bytes row = 1; public static final int ROW_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString row_; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_; /** * required bytes row = 1; */ @@ -3267,12 +3267,12 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 1; */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } private void initFields() { - row_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -3287,7 +3287,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -3303,7 +3303,7 @@ public final class IncrementCounterProcessorTestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, row_); } size += getUnknownFields().getSerializedSize(); @@ -3357,24 +3357,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom(java.io.InputStream input) @@ -3383,7 +3383,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3393,18 +3393,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3418,7 +3418,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -3426,14 +3426,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code TimeoutProcessorRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -3446,12 +3446,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -3460,7 +3460,7 @@ public final class IncrementCounterProcessorTestProtos { public Builder clear() { super.clear(); - row_ = com.google.protobuf.ByteString.EMPTY; + row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); return this; } @@ -3469,7 +3469,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorRequest_descriptor; } @@ -3499,7 +3499,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest)other); } else { @@ -3526,13 +3526,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -3545,7 +3545,7 @@ public final class IncrementCounterProcessorTestProtos { private int bitField0_; // required bytes row = 1; - private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; + private org.apache.hbase.shaded.com.google.protobuf.ByteString row_ = org.apache.hbase.shaded.com.google.protobuf.ByteString.EMPTY; /** * required bytes row = 1; */ @@ -3555,13 +3555,13 @@ public final class IncrementCounterProcessorTestProtos { /** * required bytes row = 1; */ - public com.google.protobuf.ByteString getRow() { + public org.apache.hbase.shaded.com.google.protobuf.ByteString getRow() { return row_; } /** * required bytes row = 1; */ - public Builder setRow(com.google.protobuf.ByteString value) { + public Builder setRow(org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -3592,20 +3592,20 @@ public final class IncrementCounterProcessorTestProtos { } public interface TimeoutProcessorResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code TimeoutProcessorResponse} */ public static final class TimeoutProcessorResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements TimeoutProcessorResponseOrBuilder { // Use TimeoutProcessorResponse.newBuilder() to construct. - private TimeoutProcessorResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private TimeoutProcessorResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private TimeoutProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private TimeoutProcessorResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TimeoutProcessorResponse defaultInstance; public static TimeoutProcessorResponse getDefaultInstance() { @@ -3616,19 +3616,19 @@ public final class IncrementCounterProcessorTestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TimeoutProcessorResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -3646,40 +3646,40 @@ public final class IncrementCounterProcessorTestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public TimeoutProcessorResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new TimeoutProcessorResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -3694,7 +3694,7 @@ public final class IncrementCounterProcessorTestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -3748,24 +3748,24 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom(java.io.InputStream input) @@ -3774,7 +3774,7 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3784,18 +3784,18 @@ public final class IncrementCounterProcessorTestProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3809,7 +3809,7 @@ public final class IncrementCounterProcessorTestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -3817,14 +3817,14 @@ public final class IncrementCounterProcessorTestProtos { * Protobuf type {@code TimeoutProcessorResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -3837,12 +3837,12 @@ public final class IncrementCounterProcessorTestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -3858,7 +3858,7 @@ public final class IncrementCounterProcessorTestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_TimeoutProcessorResponse_descriptor; } @@ -3881,7 +3881,7 @@ public final class IncrementCounterProcessorTestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse)other); } else { @@ -3901,13 +3901,13 @@ public final class IncrementCounterProcessorTestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -3929,52 +3929,52 @@ public final class IncrementCounterProcessorTestProtos { // @@protoc_insertion_point(class_scope:TimeoutProcessorResponse) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_IncCounterProcessorRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_IncCounterProcessorRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_IncCounterProcessorResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_IncCounterProcessorResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_FriendsOfFriendsProcessorRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_FriendsOfFriendsProcessorResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_RowSwapProcessorRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_RowSwapProcessorRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_RowSwapProcessorResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_RowSwapProcessorResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_TimeoutProcessorRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TimeoutProcessorRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_TimeoutProcessorResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TimeoutProcessorResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -3993,65 +3993,65 @@ public final class IncrementCounterProcessorTestProtos { "ratedB#IncrementCounterProcessorTestProt" + "os\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_IncCounterProcessorRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_IncCounterProcessorRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_IncCounterProcessorRequest_descriptor, new java.lang.String[] { "Row", "Counter", }); internal_static_IncCounterProcessorResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_IncCounterProcessorResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_IncCounterProcessorResponse_descriptor, new java.lang.String[] { "Response", }); internal_static_FriendsOfFriendsProcessorRequest_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FriendsOfFriendsProcessorRequest_descriptor, new java.lang.String[] { "Person", "Row", "Result", }); internal_static_FriendsOfFriendsProcessorResponse_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FriendsOfFriendsProcessorResponse_descriptor, new java.lang.String[] { "Result", }); internal_static_RowSwapProcessorRequest_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_RowSwapProcessorRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RowSwapProcessorRequest_descriptor, new java.lang.String[] { "Row1", "Row2", }); internal_static_RowSwapProcessorResponse_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_RowSwapProcessorResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RowSwapProcessorResponse_descriptor, new java.lang.String[] { }); internal_static_TimeoutProcessorRequest_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_TimeoutProcessorRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TimeoutProcessorRequest_descriptor, new java.lang.String[] { "Row", }); internal_static_TimeoutProcessorResponse_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_TimeoutProcessorResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TimeoutProcessorResponse_descriptor, new java.lang.String[] { }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/PingProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/PingProtos.java index 57903cd..b9d7e01 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/PingProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/PingProtos.java @@ -6,23 +6,23 @@ package org.apache.hadoop.hbase.coprocessor.protobuf.generated; public final class PingProtos { private PingProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface PingRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code PingRequest} */ public static final class PingRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements PingRequestOrBuilder { // Use PingRequest.newBuilder() to construct. - private PingRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private PingRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private PingRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private PingRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final PingRequest defaultInstance; public static PingRequest getDefaultInstance() { @@ -33,19 +33,19 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private PingRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -63,40 +63,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public PingRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new PingRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -111,7 +111,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -165,24 +165,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom(java.io.InputStream input) @@ -191,7 +191,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -201,18 +201,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -226,7 +226,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -234,14 +234,14 @@ public final class PingProtos { * Protobuf type {@code PingRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -254,12 +254,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -275,7 +275,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingRequest_descriptor; } @@ -298,7 +298,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest)other); } else { @@ -318,13 +318,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -347,7 +347,7 @@ public final class PingProtos { } public interface PingResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string pong = 1; /** @@ -361,21 +361,21 @@ public final class PingProtos { /** * required string pong = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getPongBytes(); } /** * Protobuf type {@code PingResponse} */ public static final class PingResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements PingResponseOrBuilder { // Use PingResponse.newBuilder() to construct. - private PingResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private PingResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private PingResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private PingResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final PingResponse defaultInstance; public static PingResponse getDefaultInstance() { @@ -386,20 +386,20 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private PingResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -422,40 +422,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public PingResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new PingResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -477,8 +477,8 @@ public final class PingProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { pong_ = s; @@ -489,17 +489,17 @@ public final class PingProtos { /** * required string pong = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getPongBytes() { java.lang.Object ref = pong_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); pong_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -519,7 +519,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -535,7 +535,7 @@ public final class PingProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getPongBytes()); } size += getUnknownFields().getSerializedSize(); @@ -589,24 +589,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom(java.io.InputStream input) @@ -615,7 +615,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -625,18 +625,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -650,7 +650,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -658,14 +658,14 @@ public final class PingProtos { * Protobuf type {@code PingResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -678,12 +678,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -701,7 +701,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_PingResponse_descriptor; } @@ -731,7 +731,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse)other); } else { @@ -760,13 +760,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -792,7 +792,7 @@ public final class PingProtos { public java.lang.String getPong() { java.lang.Object ref = pong_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); pong_ = s; return s; @@ -803,17 +803,17 @@ public final class PingProtos { /** * required string pong = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getPongBytes() { java.lang.Object ref = pong_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); pong_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -842,7 +842,7 @@ public final class PingProtos { * required string pong = 1; */ public Builder setPongBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -864,20 +864,20 @@ public final class PingProtos { } public interface CountRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code CountRequest} */ public static final class CountRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements CountRequestOrBuilder { // Use CountRequest.newBuilder() to construct. - private CountRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private CountRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private CountRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private CountRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final CountRequest defaultInstance; public static CountRequest getDefaultInstance() { @@ -888,19 +888,19 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CountRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -918,40 +918,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public CountRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new CountRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -966,7 +966,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -1020,24 +1020,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom(java.io.InputStream input) @@ -1046,7 +1046,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1056,18 +1056,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1081,7 +1081,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1089,14 +1089,14 @@ public final class PingProtos { * Protobuf type {@code CountRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1109,12 +1109,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1130,7 +1130,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountRequest_descriptor; } @@ -1153,7 +1153,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest)other); } else { @@ -1173,13 +1173,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -1202,7 +1202,7 @@ public final class PingProtos { } public interface CountResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int32 count = 1; /** @@ -1218,14 +1218,14 @@ public final class PingProtos { * Protobuf type {@code CountResponse} */ public static final class CountResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements CountResponseOrBuilder { // Use CountResponse.newBuilder() to construct. - private CountResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private CountResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private CountResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private CountResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final CountResponse defaultInstance; public static CountResponse getDefaultInstance() { @@ -1236,20 +1236,20 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CountResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1272,40 +1272,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public CountResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new CountResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -1342,7 +1342,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1358,7 +1358,7 @@ public final class PingProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(1, count_); } size += getUnknownFields().getSerializedSize(); @@ -1412,24 +1412,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom(java.io.InputStream input) @@ -1438,7 +1438,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1448,18 +1448,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1473,7 +1473,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1481,14 +1481,14 @@ public final class PingProtos { * Protobuf type {@code CountResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1501,12 +1501,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1524,7 +1524,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_CountResponse_descriptor; } @@ -1554,7 +1554,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse)other); } else { @@ -1581,13 +1581,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -1644,7 +1644,7 @@ public final class PingProtos { } public interface IncrementCountRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int32 diff = 1; /** @@ -1660,14 +1660,14 @@ public final class PingProtos { * Protobuf type {@code IncrementCountRequest} */ public static final class IncrementCountRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements IncrementCountRequestOrBuilder { // Use IncrementCountRequest.newBuilder() to construct. - private IncrementCountRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private IncrementCountRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private IncrementCountRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private IncrementCountRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final IncrementCountRequest defaultInstance; public static IncrementCountRequest getDefaultInstance() { @@ -1678,20 +1678,20 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IncrementCountRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1714,40 +1714,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public IncrementCountRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new IncrementCountRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -1784,7 +1784,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1800,7 +1800,7 @@ public final class PingProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(1, diff_); } size += getUnknownFields().getSerializedSize(); @@ -1854,24 +1854,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom(java.io.InputStream input) @@ -1880,7 +1880,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1890,18 +1890,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1915,7 +1915,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1923,14 +1923,14 @@ public final class PingProtos { * Protobuf type {@code IncrementCountRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1943,12 +1943,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1966,7 +1966,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountRequest_descriptor; } @@ -1996,7 +1996,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest)other); } else { @@ -2023,13 +2023,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -2086,7 +2086,7 @@ public final class PingProtos { } public interface IncrementCountResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required int32 count = 1; /** @@ -2102,14 +2102,14 @@ public final class PingProtos { * Protobuf type {@code IncrementCountResponse} */ public static final class IncrementCountResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements IncrementCountResponseOrBuilder { // Use IncrementCountResponse.newBuilder() to construct. - private IncrementCountResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private IncrementCountResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private IncrementCountResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private IncrementCountResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final IncrementCountResponse defaultInstance; public static IncrementCountResponse getDefaultInstance() { @@ -2120,20 +2120,20 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IncrementCountResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -2156,40 +2156,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public IncrementCountResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new IncrementCountResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -2226,7 +2226,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -2242,7 +2242,7 @@ public final class PingProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeInt32Size(1, count_); } size += getUnknownFields().getSerializedSize(); @@ -2296,24 +2296,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom(java.io.InputStream input) @@ -2322,7 +2322,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2332,18 +2332,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2357,7 +2357,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2365,14 +2365,14 @@ public final class PingProtos { * Protobuf type {@code IncrementCountResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2385,12 +2385,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -2408,7 +2408,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_IncrementCountResponse_descriptor; } @@ -2438,7 +2438,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse)other); } else { @@ -2465,13 +2465,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -2528,7 +2528,7 @@ public final class PingProtos { } public interface HelloRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional string name = 1; /** @@ -2542,21 +2542,21 @@ public final class PingProtos { /** * optional string name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes(); } /** * Protobuf type {@code HelloRequest} */ public static final class HelloRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements HelloRequestOrBuilder { // Use HelloRequest.newBuilder() to construct. - private HelloRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private HelloRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private HelloRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private HelloRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final HelloRequest defaultInstance; public static HelloRequest getDefaultInstance() { @@ -2567,20 +2567,20 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private HelloRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -2603,40 +2603,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public HelloRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new HelloRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -2658,8 +2658,8 @@ public final class PingProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { name_ = s; @@ -2670,17 +2670,17 @@ public final class PingProtos { /** * optional string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -2696,7 +2696,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -2712,7 +2712,7 @@ public final class PingProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getNameBytes()); } size += getUnknownFields().getSerializedSize(); @@ -2766,24 +2766,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom(java.io.InputStream input) @@ -2792,7 +2792,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2802,18 +2802,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -2827,7 +2827,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2835,14 +2835,14 @@ public final class PingProtos { * Protobuf type {@code HelloRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2855,12 +2855,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -2878,7 +2878,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloRequest_descriptor; } @@ -2908,7 +2908,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest)other); } else { @@ -2933,13 +2933,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -2965,7 +2965,7 @@ public final class PingProtos { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); name_ = s; return s; @@ -2976,17 +2976,17 @@ public final class PingProtos { /** * optional string name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getNameBytes() { java.lang.Object ref = name_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); name_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -3015,7 +3015,7 @@ public final class PingProtos { * optional string name = 1; */ public Builder setNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -3037,7 +3037,7 @@ public final class PingProtos { } public interface HelloResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // optional string response = 1; /** @@ -3051,21 +3051,21 @@ public final class PingProtos { /** * optional string response = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getResponseBytes(); } /** * Protobuf type {@code HelloResponse} */ public static final class HelloResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements HelloResponseOrBuilder { // Use HelloResponse.newBuilder() to construct. - private HelloResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private HelloResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private HelloResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private HelloResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final HelloResponse defaultInstance; public static HelloResponse getDefaultInstance() { @@ -3076,20 +3076,20 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private HelloResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -3112,40 +3112,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public HelloResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new HelloResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -3167,8 +3167,8 @@ public final class PingProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { response_ = s; @@ -3179,17 +3179,17 @@ public final class PingProtos { /** * optional string response = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getResponseBytes() { java.lang.Object ref = response_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); response_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -3205,7 +3205,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -3221,7 +3221,7 @@ public final class PingProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getResponseBytes()); } size += getUnknownFields().getSerializedSize(); @@ -3275,24 +3275,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom(java.io.InputStream input) @@ -3301,7 +3301,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3311,18 +3311,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3336,7 +3336,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -3344,14 +3344,14 @@ public final class PingProtos { * Protobuf type {@code HelloResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -3364,12 +3364,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -3387,7 +3387,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_HelloResponse_descriptor; } @@ -3417,7 +3417,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse)other); } else { @@ -3442,13 +3442,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -3474,7 +3474,7 @@ public final class PingProtos { public java.lang.String getResponse() { java.lang.Object ref = response_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); response_ = s; return s; @@ -3485,17 +3485,17 @@ public final class PingProtos { /** * optional string response = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getResponseBytes() { java.lang.Object ref = response_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); response_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -3524,7 +3524,7 @@ public final class PingProtos { * optional string response = 1; */ public Builder setResponseBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -3546,20 +3546,20 @@ public final class PingProtos { } public interface NoopRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code NoopRequest} */ public static final class NoopRequest extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements NoopRequestOrBuilder { // Use NoopRequest.newBuilder() to construct. - private NoopRequest(com.google.protobuf.GeneratedMessage.Builder builder) { + private NoopRequest(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private NoopRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private NoopRequest(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final NoopRequest defaultInstance; public static NoopRequest getDefaultInstance() { @@ -3570,19 +3570,19 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private NoopRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -3600,40 +3600,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public NoopRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new NoopRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -3648,7 +3648,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -3702,24 +3702,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom(java.io.InputStream input) @@ -3728,7 +3728,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3738,18 +3738,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -3763,7 +3763,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -3771,14 +3771,14 @@ public final class PingProtos { * Protobuf type {@code NoopRequest} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopRequest_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -3791,12 +3791,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -3812,7 +3812,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopRequest_descriptor; } @@ -3835,7 +3835,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest)other); } else { @@ -3855,13 +3855,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest) e.getUnfinishedMessage(); throw e; } finally { @@ -3884,20 +3884,20 @@ public final class PingProtos { } public interface NoopResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code NoopResponse} */ public static final class NoopResponse extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements NoopResponseOrBuilder { // Use NoopResponse.newBuilder() to construct. - private NoopResponse(com.google.protobuf.GeneratedMessage.Builder builder) { + private NoopResponse(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private NoopResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private NoopResponse(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final NoopResponse defaultInstance; public static NoopResponse getDefaultInstance() { @@ -3908,19 +3908,19 @@ public final class PingProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private NoopResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -3938,40 +3938,40 @@ public final class PingProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public NoopResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new NoopResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -3986,7 +3986,7 @@ public final class PingProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -4040,24 +4040,24 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom(java.io.InputStream input) @@ -4066,7 +4066,7 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -4076,18 +4076,18 @@ public final class PingProtos { } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -4101,7 +4101,7 @@ public final class PingProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -4109,14 +4109,14 @@ public final class PingProtos { * Protobuf type {@code NoopResponse} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopResponse_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -4129,12 +4129,12 @@ public final class PingProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -4150,7 +4150,7 @@ public final class PingProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.internal_static_NoopResponse_descriptor; } @@ -4173,7 +4173,7 @@ public final class PingProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse) { return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse)other); } else { @@ -4193,13 +4193,13 @@ public final class PingProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse) e.getUnfinishedMessage(); throw e; } finally { @@ -4225,7 +4225,7 @@ public final class PingProtos { * Protobuf service {@code PingService} */ public static abstract class PingService - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected PingService() {} public interface Interface { @@ -4233,103 +4233,103 @@ public final class PingProtos { * rpc ping(.PingRequest) returns (.PingResponse); */ public abstract void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc count(.CountRequest) returns (.CountResponse); */ public abstract void count( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc increment(.IncrementCountRequest) returns (.IncrementCountResponse); */ public abstract void increment( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc hello(.HelloRequest) returns (.HelloResponse); */ public abstract void hello( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc noop(.NoopRequest) returns (.NoopResponse); */ public abstract void noop( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new PingService() { @java.lang.Override public void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.ping(controller, request, done); } @java.lang.Override public void count( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.count(controller, request, done); } @java.lang.Override public void increment( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.increment(controller, request, done); } @java.lang.Override public void hello( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.hello(controller, request, done); } @java.lang.Override public void noop( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.noop(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -4351,9 +4351,9 @@ public final class PingProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -4375,9 +4375,9 @@ public final class PingProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -4406,58 +4406,58 @@ public final class PingProtos { * rpc ping(.PingRequest) returns (.PingResponse); */ public abstract void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc count(.CountRequest) returns (.CountResponse); */ public abstract void count( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc increment(.IncrementCountRequest) returns (.IncrementCountResponse); */ public abstract void increment( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc hello(.HelloRequest) returns (.HelloResponse); */ public abstract void hello( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc noop(.NoopRequest) returns (.NoopResponse); */ public abstract void noop( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -4466,27 +4466,27 @@ public final class PingProtos { switch(method.getIndex()) { case 0: this.ping(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.count(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 2: this.increment(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 3: this.hello(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 4: this.noop(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -4494,9 +4494,9 @@ public final class PingProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -4518,9 +4518,9 @@ public final class PingProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -4542,92 +4542,93 @@ public final class PingProtos { } } - public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + public static Stub newStub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } - public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingService implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + public static final class Stub + extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingService + implements Interface { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse.getDefaultInstance())); } public void count( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse.getDefaultInstance())); } public void increment( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse.getDefaultInstance())); } public void hello( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse.getDefaultInstance())); } public void noop( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse.getDefaultInstance())); @@ -4635,48 +4636,48 @@ public final class PingProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse count( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse increment( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse hello( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse noop( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.PingResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -4686,9 +4687,9 @@ public final class PingProtos { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse count( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.CountResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, @@ -4698,9 +4699,9 @@ public final class PingProtos { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse increment( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.IncrementCountResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, @@ -4710,9 +4711,9 @@ public final class PingProtos { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse hello( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.HelloResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(3), controller, @@ -4722,9 +4723,9 @@ public final class PingProtos { public org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse noop( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopRequest request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.PingProtos.NoopResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(4), controller, @@ -4737,62 +4738,62 @@ public final class PingProtos { // @@protoc_insertion_point(class_scope:PingService) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_PingRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_PingRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_PingResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_PingResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_CountRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_CountRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_CountResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_CountResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_IncrementCountRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_IncrementCountRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_IncrementCountResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_IncrementCountResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_HelloRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_HelloRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_HelloResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_HelloResponse_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_NoopRequest_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_NoopRequest_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_NoopResponse_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_NoopResponse_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -4812,77 +4813,77 @@ public final class PingProtos { "eBJ\n6org.apache.hadoop.hbase.coprocessor" + ".protobuf.generatedB\nPingProtos\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_PingRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_PingRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_PingRequest_descriptor, new java.lang.String[] { }); internal_static_PingResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_PingResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_PingResponse_descriptor, new java.lang.String[] { "Pong", }); internal_static_CountRequest_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_CountRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_CountRequest_descriptor, new java.lang.String[] { }); internal_static_CountResponse_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_CountResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_CountResponse_descriptor, new java.lang.String[] { "Count", }); internal_static_IncrementCountRequest_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_IncrementCountRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_IncrementCountRequest_descriptor, new java.lang.String[] { "Diff", }); internal_static_IncrementCountResponse_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_IncrementCountResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_IncrementCountResponse_descriptor, new java.lang.String[] { "Count", }); internal_static_HelloRequest_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_HelloRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_HelloRequest_descriptor, new java.lang.String[] { "Name", }); internal_static_HelloResponse_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_HelloResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_HelloResponse_descriptor, new java.lang.String[] { "Response", }); internal_static_NoopRequest_descriptor = getDescriptor().getMessageTypes().get(8); internal_static_NoopRequest_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_NoopRequest_descriptor, new java.lang.String[] { }); internal_static_NoopResponse_descriptor = getDescriptor().getMessageTypes().get(9); internal_static_NoopResponse_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_NoopResponse_descriptor, new java.lang.String[] { }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/errorhandling/TestForeignExceptionSerialization.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/errorhandling/TestForeignExceptionSerialization.java index f893555..40abb49 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/errorhandling/TestForeignExceptionSerialization.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/errorhandling/TestForeignExceptionSerialization.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; /** * Test that we correctly serialize exceptions from a remote source diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/AbstractTestIPC.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/AbstractTestIPC.java index 69c8fe2..981519f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/AbstractTestIPC.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/AbstractTestIPC.java @@ -62,12 +62,12 @@ import org.junit.Test; import com.google.common.collect.ImmutableList; import com.google.common.collect.Lists; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.BlockingService; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Some basic ipc tests. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestAsyncIPC.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestAsyncIPC.java index b9d390a..4df3dc8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestAsyncIPC.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestAsyncIPC.java @@ -67,11 +67,11 @@ import org.junit.runner.RunWith; import org.junit.runners.Parameterized; import org.junit.runners.Parameterized.Parameters; -import com.google.protobuf.ByteString; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcChannel; @RunWith(Parameterized.class) @Category({ RPCTests.class, SmallTests.class }) diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestIPC.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestIPC.java index 3fc1259..1422cb3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestIPC.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestIPC.java @@ -59,8 +59,8 @@ import org.mockito.Mockito; import org.mockito.invocation.InvocationOnMock; import org.mockito.stubbing.Answer; -import com.google.protobuf.ByteString; -import com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; @Category({ RPCTests.class, SmallTests.class }) public class TestIPC extends AbstractTestIPC { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestProtoBufRpc.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestProtoBufRpc.java index 81869b4..8820ec7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestProtoBufRpc.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestProtoBufRpc.java @@ -42,10 +42,10 @@ import org.junit.Before; import org.junit.After; import org.junit.experimental.categories.Category; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.BlockingService; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test for testing protocol buffer based RPC mechanism. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcHandlerException.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcHandlerException.java index a37ba11..256ee8e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcHandlerException.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcHandlerException.java @@ -19,11 +19,11 @@ package org.apache.hadoop.hbase.ipc; import com.google.common.collect.ImmutableList; import com.google.common.collect.Lists; -import com.google.protobuf.BlockingService; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestSimpleRpcScheduler.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestSimpleRpcScheduler.java index 97ef973..a076c85 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestSimpleRpcScheduler.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestSimpleRpcScheduler.java @@ -73,7 +73,7 @@ import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableMap; import com.google.common.collect.ImmutableSet; import com.google.common.collect.Maps; -import com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Message; @Category({RPCTests.class, SmallTests.class}) public class TestSimpleRpcScheduler { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProcedureProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProcedureProtos.java index f065550..d407b7d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProcedureProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProcedureProtos.java @@ -6,10 +6,10 @@ package org.apache.hadoop.hbase.ipc.protobuf.generated; public final class TestProcedureProtos { private TestProcedureProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface TestTableDDLStateDataOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string table_name = 1; /** @@ -23,21 +23,21 @@ public final class TestProcedureProtos { /** * required string table_name = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getTableNameBytes(); } /** * Protobuf type {@code TestTableDDLStateData} */ public static final class TestTableDDLStateData extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements TestTableDDLStateDataOrBuilder { // Use TestTableDDLStateData.newBuilder() to construct. - private TestTableDDLStateData(com.google.protobuf.GeneratedMessage.Builder builder) { + private TestTableDDLStateData(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private TestTableDDLStateData(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private TestTableDDLStateData(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TestTableDDLStateData defaultInstance; public static TestTableDDLStateData getDefaultInstance() { @@ -48,20 +48,20 @@ public final class TestProcedureProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TestTableDDLStateData( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -84,40 +84,40 @@ public final class TestProcedureProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public TestTableDDLStateData parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new TestTableDDLStateData(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -139,8 +139,8 @@ public final class TestProcedureProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { tableName_ = s; @@ -151,17 +151,17 @@ public final class TestProcedureProtos { /** * required string table_name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getTableNameBytes() { java.lang.Object ref = tableName_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); tableName_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -181,7 +181,7 @@ public final class TestProcedureProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -197,7 +197,7 @@ public final class TestProcedureProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getTableNameBytes()); } size += getUnknownFields().getSerializedSize(); @@ -213,24 +213,24 @@ public final class TestProcedureProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(java.io.InputStream input) @@ -239,7 +239,7 @@ public final class TestProcedureProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -249,18 +249,18 @@ public final class TestProcedureProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -274,7 +274,7 @@ public final class TestProcedureProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -282,14 +282,14 @@ public final class TestProcedureProtos { * Protobuf type {@code TestTableDDLStateData} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateDataOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -302,12 +302,12 @@ public final class TestProcedureProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -325,7 +325,7 @@ public final class TestProcedureProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor; } @@ -355,7 +355,7 @@ public final class TestProcedureProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) { return mergeFrom((org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData)other); } else { @@ -384,13 +384,13 @@ public final class TestProcedureProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) e.getUnfinishedMessage(); throw e; } finally { @@ -416,7 +416,7 @@ public final class TestProcedureProtos { public java.lang.String getTableName() { java.lang.Object ref = tableName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); tableName_ = s; return s; @@ -427,17 +427,17 @@ public final class TestProcedureProtos { /** * required string table_name = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getTableNameBytes() { java.lang.Object ref = tableName_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); tableName_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -466,7 +466,7 @@ public final class TestProcedureProtos { * required string table_name = 1; */ public Builder setTableNameBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -487,17 +487,17 @@ public final class TestProcedureProtos { // @@protoc_insertion_point(class_scope:TestTableDDLStateData) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_TestTableDDLStateData_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TestTableDDLStateData_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -506,23 +506,23 @@ public final class TestProcedureProtos { ".hadoop.hbase.ipc.protobuf.generatedB\023Te" + "stProcedureProtos\210\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_TestTableDDLStateData_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_TestTableDDLStateData_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TestTableDDLStateData_descriptor, new java.lang.String[] { "TableName", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProtos.java index 3da0254..d0947ff 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestProtos.java @@ -6,23 +6,23 @@ package org.apache.hadoop.hbase.ipc.protobuf.generated; public final class TestProtos { private TestProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } public interface EmptyRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code EmptyRequestProto} */ public static final class EmptyRequestProto extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements EmptyRequestProtoOrBuilder { // Use EmptyRequestProto.newBuilder() to construct. - private EmptyRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EmptyRequestProto(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private EmptyRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private EmptyRequestProto(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final EmptyRequestProto defaultInstance; public static EmptyRequestProto getDefaultInstance() { @@ -33,19 +33,19 @@ public final class TestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EmptyRequestProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -63,40 +63,40 @@ public final class TestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public EmptyRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new EmptyRequestProto(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -111,7 +111,7 @@ public final class TestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -165,24 +165,24 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(java.io.InputStream input) @@ -191,7 +191,7 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -201,18 +201,18 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -226,7 +226,7 @@ public final class TestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -234,14 +234,14 @@ public final class TestProtos { * Protobuf type {@code EmptyRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -254,12 +254,12 @@ public final class TestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -275,7 +275,7 @@ public final class TestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor; } @@ -298,7 +298,7 @@ public final class TestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto) { return mergeFrom((org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)other); } else { @@ -318,13 +318,13 @@ public final class TestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto) e.getUnfinishedMessage(); throw e; } finally { @@ -347,20 +347,20 @@ public final class TestProtos { } public interface EmptyResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code EmptyResponseProto} */ public static final class EmptyResponseProto extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements EmptyResponseProtoOrBuilder { // Use EmptyResponseProto.newBuilder() to construct. - private EmptyResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EmptyResponseProto(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private EmptyResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private EmptyResponseProto(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final EmptyResponseProto defaultInstance; public static EmptyResponseProto getDefaultInstance() { @@ -371,19 +371,19 @@ public final class TestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EmptyResponseProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -401,40 +401,40 @@ public final class TestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public EmptyResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new EmptyResponseProto(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -449,7 +449,7 @@ public final class TestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); @@ -503,24 +503,24 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(java.io.InputStream input) @@ -529,7 +529,7 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -539,18 +539,18 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -564,7 +564,7 @@ public final class TestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -572,14 +572,14 @@ public final class TestProtos { * Protobuf type {@code EmptyResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -592,12 +592,12 @@ public final class TestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -613,7 +613,7 @@ public final class TestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor; } @@ -636,7 +636,7 @@ public final class TestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto) { return mergeFrom((org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto)other); } else { @@ -656,13 +656,13 @@ public final class TestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto) e.getUnfinishedMessage(); throw e; } finally { @@ -685,7 +685,7 @@ public final class TestProtos { } public interface EchoRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string message = 1; /** @@ -699,21 +699,21 @@ public final class TestProtos { /** * required string message = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getMessageBytes(); } /** * Protobuf type {@code EchoRequestProto} */ public static final class EchoRequestProto extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements EchoRequestProtoOrBuilder { // Use EchoRequestProto.newBuilder() to construct. - private EchoRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EchoRequestProto(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private EchoRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private EchoRequestProto(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final EchoRequestProto defaultInstance; public static EchoRequestProto getDefaultInstance() { @@ -724,20 +724,20 @@ public final class TestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EchoRequestProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -760,40 +760,40 @@ public final class TestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public EchoRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new EchoRequestProto(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -815,8 +815,8 @@ public final class TestProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { message_ = s; @@ -827,17 +827,17 @@ public final class TestProtos { /** * required string message = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getMessageBytes() { java.lang.Object ref = message_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); message_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -857,7 +857,7 @@ public final class TestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -873,7 +873,7 @@ public final class TestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getMessageBytes()); } size += getUnknownFields().getSerializedSize(); @@ -927,24 +927,24 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(java.io.InputStream input) @@ -953,7 +953,7 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -963,18 +963,18 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -988,7 +988,7 @@ public final class TestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -996,14 +996,14 @@ public final class TestProtos { * Protobuf type {@code EchoRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1016,12 +1016,12 @@ public final class TestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1039,7 +1039,7 @@ public final class TestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor; } @@ -1069,7 +1069,7 @@ public final class TestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto) { return mergeFrom((org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto)other); } else { @@ -1098,13 +1098,13 @@ public final class TestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto) e.getUnfinishedMessage(); throw e; } finally { @@ -1130,7 +1130,7 @@ public final class TestProtos { public java.lang.String getMessage() { java.lang.Object ref = message_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); message_ = s; return s; @@ -1141,17 +1141,17 @@ public final class TestProtos { /** * required string message = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getMessageBytes() { java.lang.Object ref = message_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); message_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1180,7 +1180,7 @@ public final class TestProtos { * required string message = 1; */ public Builder setMessageBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1202,7 +1202,7 @@ public final class TestProtos { } public interface EchoResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + extends org.apache.hbase.shaded.com.google.protobuf.MessageOrBuilder { // required string message = 1; /** @@ -1216,21 +1216,21 @@ public final class TestProtos { /** * required string message = 1; */ - com.google.protobuf.ByteString + org.apache.hbase.shaded.com.google.protobuf.ByteString getMessageBytes(); } /** * Protobuf type {@code EchoResponseProto} */ public static final class EchoResponseProto extends - com.google.protobuf.GeneratedMessage + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage implements EchoResponseProtoOrBuilder { // Use EchoResponseProto.newBuilder() to construct. - private EchoResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EchoResponseProto(org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private EchoResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private EchoResponseProto(boolean noInit) { this.unknownFields = org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final EchoResponseProto defaultInstance; public static EchoResponseProto getDefaultInstance() { @@ -1241,20 +1241,20 @@ public final class TestProtos { return defaultInstance; } - private final com.google.protobuf.UnknownFieldSet unknownFields; + private final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override - public final com.google.protobuf.UnknownFieldSet + public final org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EchoResponseProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = + org.apache.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { @@ -1277,40 +1277,40 @@ public final class TestProtos { } } } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( + throw new org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { + public static org.apache.hbase.shaded.com.google.protobuf.Parser PARSER = + new org.apache.hbase.shaded.com.google.protobuf.AbstractParser() { public EchoResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return new EchoResponseProto(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public org.apache.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } @@ -1332,8 +1332,8 @@ public final class TestProtos { if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; + org.apache.hbase.shaded.com.google.protobuf.ByteString bs = + (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { message_ = s; @@ -1344,17 +1344,17 @@ public final class TestProtos { /** * required string message = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getMessageBytes() { java.lang.Object ref = message_; if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); message_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } @@ -1374,7 +1374,7 @@ public final class TestProtos { return true; } - public void writeTo(com.google.protobuf.CodedOutputStream output) + public void writeTo(org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1390,7 +1390,7 @@ public final class TestProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream + size += org.apache.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBytesSize(1, getMessageBytes()); } size += getUnknownFields().getSerializedSize(); @@ -1444,24 +1444,24 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ByteString data, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom( byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(java.io.InputStream input) @@ -1470,7 +1470,7 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1480,18 +1480,18 @@ public final class TestProtos { } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseDelimitedFrom( java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom( - com.google.protobuf.CodedInputStream input) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } @@ -1505,7 +1505,7 @@ public final class TestProtos { @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1513,14 +1513,14 @@ public final class TestProtos { * Protobuf type {@code EchoResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.Builder implements org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor + public static final org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1533,12 +1533,12 @@ public final class TestProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { @@ -1556,7 +1556,7 @@ public final class TestProtos { return create().mergeFrom(buildPartial()); } - public com.google.protobuf.Descriptors.Descriptor + public org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor; } @@ -1586,7 +1586,7 @@ public final class TestProtos { return result; } - public Builder mergeFrom(com.google.protobuf.Message other) { + public Builder mergeFrom(org.apache.hbase.shaded.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto) { return mergeFrom((org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto)other); } else { @@ -1615,13 +1615,13 @@ public final class TestProtos { } public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) + org.apache.hbase.shaded.com.google.protobuf.CodedInputStream input, + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { + } catch (org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto) e.getUnfinishedMessage(); throw e; } finally { @@ -1647,7 +1647,7 @@ public final class TestProtos { public java.lang.String getMessage() { java.lang.Object ref = message_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) + java.lang.String s = ((org.apache.hbase.shaded.com.google.protobuf.ByteString) ref) .toStringUtf8(); message_ = s; return s; @@ -1658,17 +1658,17 @@ public final class TestProtos { /** * required string message = 1; */ - public com.google.protobuf.ByteString + public org.apache.hbase.shaded.com.google.protobuf.ByteString getMessageBytes() { java.lang.Object ref = message_; if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( + org.apache.hbase.shaded.com.google.protobuf.ByteString b = + org.apache.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); message_ = b; return b; } else { - return (com.google.protobuf.ByteString) ref; + return (org.apache.hbase.shaded.com.google.protobuf.ByteString) ref; } } /** @@ -1697,7 +1697,7 @@ public final class TestProtos { * required string message = 1; */ public Builder setMessageBytes( - com.google.protobuf.ByteString value) { + org.apache.hbase.shaded.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } @@ -1718,32 +1718,32 @@ public final class TestProtos { // @@protoc_insertion_point(class_scope:EchoResponseProto) } - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_EmptyRequestProto_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_EmptyRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_EmptyResponseProto_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_EmptyResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_EchoRequestProto_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_EchoRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_EchoResponseProto_descriptor; private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_EchoResponseProto_fieldAccessorTable; - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1753,41 +1753,41 @@ public final class TestProtos { "sage\030\001 \002(\tB?\n.org.apache.hadoop.hbase.ip" + "c.protobuf.generatedB\nTestProtos\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_EmptyRequestProto_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_EmptyRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_EmptyRequestProto_descriptor, new java.lang.String[] { }); internal_static_EmptyResponseProto_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_EmptyResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_EmptyResponseProto_descriptor, new java.lang.String[] { }); internal_static_EchoRequestProto_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_EchoRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_EchoRequestProto_descriptor, new java.lang.String[] { "Message", }); internal_static_EchoResponseProto_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_EchoResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( + org.apache.hbase.shaded.com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_EchoResponseProto_descriptor, new java.lang.String[] { "Message", }); return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java index 3fd34e9..31fcd63 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java @@ -6,7 +6,7 @@ package org.apache.hadoop.hbase.ipc.protobuf.generated; public final class TestRpcServiceProtos { private TestRpcServiceProtos() {} public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { + org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { } /** * Protobuf service {@code TestProtobufRpcProto} @@ -17,7 +17,7 @@ public final class TestRpcServiceProtos { * */ public static abstract class TestProtobufRpcProto - implements com.google.protobuf.Service { + implements org.apache.hbase.shaded.com.google.protobuf.Service { protected TestProtobufRpcProto() {} public interface Interface { @@ -25,71 +25,71 @@ public final class TestRpcServiceProtos { * rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto); */ public abstract void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc echo(.EchoRequestProto) returns (.EchoResponseProto); */ public abstract void echo( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc error(.EmptyRequestProto) returns (.EmptyResponseProto); */ public abstract void error( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); } - public static com.google.protobuf.Service newReflectiveService( + public static org.apache.hbase.shaded.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new TestProtobufRpcProto() { @java.lang.Override public void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.ping(controller, request, done); } @java.lang.Override public void echo( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.echo(controller, request, done); } @java.lang.Override public void error( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { impl.error(controller, request, done); } }; } - public static com.google.protobuf.BlockingService + public static org.apache.hbase.shaded.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor + return new org.apache.hbase.shaded.com.google.protobuf.BlockingService() { + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { + public final org.apache.hbase.shaded.com.google.protobuf.Message callBlockingMethod( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request) + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + @@ -107,9 +107,9 @@ public final class TestRpcServiceProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -127,9 +127,9 @@ public final class TestRpcServiceProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -154,42 +154,42 @@ public final class TestRpcServiceProtos { * rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto); */ public abstract void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc echo(.EchoRequestProto) returns (.EchoResponseProto); */ public abstract void echo( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); /** * rpc error(.EmptyRequestProto) returns (.EmptyResponseProto); */ public abstract void error( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done); + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done); public static final - com.google.protobuf.Descriptors.ServiceDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.ipc.protobuf.generated.TestRpcServiceProtos.getDescriptor().getServices().get(0); } - public final com.google.protobuf.Descriptors.ServiceDescriptor + public final org.apache.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.Message request, + org.apache.hbase.shaded.com.google.protobuf.RpcCallback< + org.apache.hbase.shaded.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + @@ -198,17 +198,17 @@ public final class TestRpcServiceProtos { switch(method.getIndex()) { case 0: this.ping(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.echo(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 2: this.error(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request, - com.google.protobuf.RpcUtil.specializeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: @@ -216,9 +216,9 @@ public final class TestRpcServiceProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + @@ -236,9 +236,9 @@ public final class TestRpcServiceProtos { } } - public final com.google.protobuf.Message + public final org.apache.hbase.shaded.com.google.protobuf.Message getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + @@ -257,61 +257,61 @@ public final class TestRpcServiceProtos { } public static Stub newStub( - com.google.protobuf.RpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hbase.ipc.protobuf.generated.TestRpcServiceProtos.TestProtobufRpcProto implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { + private Stub(org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.RpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.RpcChannel channel; - public com.google.protobuf.RpcChannel getChannel() { + public org.apache.hbase.shaded.com.google.protobuf.RpcChannel getChannel() { return channel; } public void ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance())); } public void echo( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance())); } public void error( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request, - com.google.protobuf.RpcCallback done) { + org.apache.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( + org.apache.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance())); @@ -319,38 +319,38 @@ public final class TestRpcServiceProtos { } public static BlockingInterface newBlockingStub( - com.google.protobuf.BlockingRpcChannel channel) { + org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto echo( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto error( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request) - throws com.google.protobuf.ServiceException; + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { - private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { + private BlockingStub(org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } - private final com.google.protobuf.BlockingRpcChannel channel; + private final org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, @@ -360,9 +360,9 @@ public final class TestRpcServiceProtos { public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto echo( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, @@ -372,9 +372,9 @@ public final class TestRpcServiceProtos { public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto error( - com.google.protobuf.RpcController controller, + org.apache.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request) - throws com.google.protobuf.ServiceException { + throws org.apache.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, @@ -388,11 +388,11 @@ public final class TestRpcServiceProtos { } - public static com.google.protobuf.Descriptors.FileDescriptor + public static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -405,17 +405,17 @@ public final class TestRpcServiceProtos { "rotobuf.generatedB\024TestRpcServiceProtos\210" + "\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public org.apache.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; return null; } }; - com.google.protobuf.Descriptors.FileDescriptor + org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { + new org.apache.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.getDescriptor(), }, assigner); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java index 32e3058..9c7c1f7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java @@ -71,8 +71,8 @@ import org.junit.experimental.categories.Category; import org.mockito.Mockito; import com.google.common.collect.Multimap; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test cases for the atomic load error handling of the bulk load functionality. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java index 69f2e35..146b12a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java @@ -111,9 +111,9 @@ import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * A mock RegionServer implementation. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestCatalogJanitor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestCatalogJanitor.java index e417c8e..d579138 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestCatalogJanitor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestCatalogJanitor.java @@ -94,9 +94,9 @@ import org.mockito.Mockito; import org.mockito.invocation.InvocationOnMock; import org.mockito.stubbing.Answer; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({MasterTests.class, SmallTests.class}) public class TestCatalogJanitor { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestHMasterRPCException.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestHMasterRPCException.java index 37d6940..733ac29 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestHMasterRPCException.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestHMasterRPCException.java @@ -48,8 +48,8 @@ import org.junit.Before; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({ MasterTests.class, MediumTests.class }) public class TestHMasterRPCException { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterNoCluster.java index cafee7a..64acf2a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterNoCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterNoCluster.java @@ -69,7 +69,7 @@ import org.junit.experimental.categories.Category; import org.junit.rules.TestRule; import org.mockito.Mockito; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Standup the master and fake it to test various aspects of master function. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java index b95089b..c231101 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java @@ -69,7 +69,7 @@ import org.junit.experimental.categories.Category; import org.mockito.Mockito; import com.google.common.collect.Lists; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Test the master-related aspects of a snapshot diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java index 9d171f0..3e91b09 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java @@ -38,8 +38,8 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.mockito.Mockito; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import java.util.ArrayList; import java.util.HashMap; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java index b2d8b38..a3b5b23 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java @@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Class to test ProtobufUtil. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/OOMERegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/OOMERegionServer.java index 35258f2..b43d2de 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/OOMERegionServer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/OOMERegionServer.java @@ -30,7 +30,7 @@ import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.RequestConverter; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutateRequest; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * A region server that will OOME. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/RegionAsTable.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/RegionAsTable.java index f65bc5d..c715f5d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/RegionAsTable.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/RegionAsTable.java @@ -45,10 +45,10 @@ import org.apache.hadoop.hbase.client.coprocessor.Batch.Callback; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -import com.google.protobuf.Descriptors.MethodDescriptor; -import com.google.protobuf.Message; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hbase.shaded.com.google.protobuf.Message; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * An implementation of {@link Table} that sits directly on a Region; it decorates the passed in diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java index 1c1a603..214f62e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java @@ -71,7 +71,7 @@ import org.junit.experimental.categories.Category; import com.google.common.collect.Iterators; import com.google.common.collect.Sets; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category(LargeTests.class) public class TestEndToEndSplitTransaction { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java index ed7623c..a2c5e72 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import com.google.common.collect.ImmutableList; import com.google.common.collect.Lists; import com.google.common.collect.Maps; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; import org.apache.commons.lang.RandomStringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java index 5fde726..51d4eca 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java @@ -46,7 +46,7 @@ import org.junit.Assert; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({RegionServerTests.class, SmallTests.class}) public class TestHRegionInfo { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java index 9183e18..4134f4b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java @@ -98,7 +98,7 @@ import org.junit.experimental.categories.Category; import org.junit.rules.TestName; import com.google.common.collect.Lists; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Tests of HRegion methods for replaying flush, compaction, region open, etc events for secondary diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java index edad059..3350812 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java @@ -47,7 +47,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.mockito.Mockito; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Tests that verify certain RPCs get a higher QoS. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRSStatusServlet.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRSStatusServlet.java index a9115f3..c95f29d2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRSStatusServlet.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRSStatusServlet.java @@ -43,8 +43,8 @@ import org.junit.experimental.categories.Category; import org.mockito.Mockito; import com.google.common.collect.Lists; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.HConstants; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java index cd4410f..47f9dfb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java @@ -82,8 +82,8 @@ import org.junit.rules.TestName; import org.junit.rules.TestRule; import com.google.common.base.Joiner; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Like {@link TestRegionMergeTransaction} in that we're testing diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java index 99f5801..460f1c2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java @@ -59,7 +59,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Tests for region replicas. Sad that we cannot isolate these without bringing up a whole diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java index 0ee75a8..336f91e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java @@ -51,7 +51,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Tests on the region server, without the master. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerHeartbeatMessages.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerHeartbeatMessages.java index 1935c0a..f132231 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerHeartbeatMessages.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerHeartbeatMessages.java @@ -66,8 +66,8 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Here we test to make sure that scans return the expected Results when the server is sending the diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java index 470c6d1..85ef043 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java @@ -63,10 +63,10 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({RegionServerTests.class, MediumTests.class}) public class TestServerCustomProtocol { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java index fe620e7..106328a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java @@ -106,8 +106,8 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Like TestSplitTransaction in that we're testing {@link SplitTransactionImpl} diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java index 775e4a5..3fef678 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java @@ -75,7 +75,7 @@ import org.junit.Before; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({ReplicationTests.class, LargeTests.class}) public class TestMasterReplication { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/AbstractTestSecureIPC.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/AbstractTestSecureIPC.java index 7e99cc0..b01d93c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/AbstractTestSecureIPC.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/AbstractTestSecureIPC.java @@ -34,8 +34,8 @@ import java.util.List; import java.util.Properties; import java.util.concurrent.ThreadLocalRandom; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.CommonConfigurationKeys; import org.apache.hadoop.hbase.Cell; @@ -64,8 +64,8 @@ import org.junit.rules.ExpectedException; import org.mockito.Mockito; import com.google.common.collect.Lists; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; import javax.security.sasl.SaslException; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java index 22a9748..fc60838 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java @@ -66,8 +66,8 @@ import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread; import com.google.common.collect.Lists; import com.google.common.collect.Maps; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Utility methods for testing security diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java index 37c42a0..e7f1ad6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java @@ -125,11 +125,11 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.RpcCallback; -import com.google.protobuf.RpcController; -import com.google.protobuf.Service; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.RpcCallback; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.Service; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Performs authorization checks for common operations, according to different diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java index d5834fd..4625613 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java @@ -55,7 +55,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import com.google.common.collect.ListMultimap; -import com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; @Category({SecurityTests.class, MediumTests.class}) public class TestNamespaceCommands extends SecureTestUtil { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestGenerateDelegationToken.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestGenerateDelegationToken.java index e3d3c27..041dec3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestGenerateDelegationToken.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestGenerateDelegationToken.java @@ -62,7 +62,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({ SecurityTests.class, MediumTests.class }) public class TestGenerateDelegationToken { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java index faac8eb..4daa607 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java @@ -85,10 +85,10 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.BlockingRpcChannel; -import com.google.protobuf.BlockingService; -import com.google.protobuf.RpcController; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.BlockingRpcChannel; +import org.apache.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hbase.shaded.com.google.protobuf.RpcController; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Tests for authentication token creation and usage diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java index ab2bacc..ab08c0b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java @@ -72,7 +72,7 @@ import org.junit.Rule; import org.junit.Test; import org.junit.rules.TestName; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; /** * Base test class for visibility labels basic features diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsOpWithDifferentUsersNoACL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsOpWithDifferentUsersNoACL.java index 06fc7be..28b3378 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsOpWithDifferentUsersNoACL.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsOpWithDifferentUsersNoACL.java @@ -43,7 +43,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.junit.rules.TestName; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({SecurityTests.class, MediumTests.class}) public class TestVisibilityLabelsOpWithDifferentUsersNoACL { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java index f67296d..9de20c7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java @@ -56,7 +56,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.junit.rules.TestName; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({SecurityTests.class, MediumTests.class}) public class TestVisibilityLabelsWithACL { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDefaultVisLabelService.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDefaultVisLabelService.java index a229bdb..d8f032d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDefaultVisLabelService.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDefaultVisLabelService.java @@ -55,7 +55,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({SecurityTests.class, MediumTests.class}) public class TestVisibilityLabelsWithDefaultVisLabelService extends TestVisibilityLabels { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java index 1410c78..66e40da 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java @@ -54,7 +54,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.junit.rules.TestName; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({SecurityTests.class, MediumTests.class}) public class TestVisibilityLablesWithGroups { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestWithDisabledAuthorization.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestWithDisabledAuthorization.java index 1d027d0..56f08fb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestWithDisabledAuthorization.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestWithDisabledAuthorization.java @@ -48,7 +48,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import org.junit.rules.TestName; -import com.google.protobuf.ByteString; +import org.apache.hbase.shaded.com.google.protobuf.ByteString; @Category({SecurityTests.class, LargeTests.class}) public class TestWithDisabledAuthorization { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java index 666eea3..2c4cb0e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java @@ -73,7 +73,7 @@ import org.apache.hadoop.hbase.util.FSUtils; import org.apache.hadoop.hbase.util.MD5Hash; import org.junit.Assert; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; /** * Utilities class for snapshots diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java index 835f92e..7476586 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.snapshot; -import com.google.protobuf.InvalidProtocolBufferException; +import org.apache.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java index bde3e49..5f2eece 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java @@ -46,7 +46,7 @@ import org.junit.Test; import org.junit.experimental.categories.Category; import com.google.common.collect.Lists; -import com.google.protobuf.ServiceException; +import org.apache.hbase.shaded.com.google.protobuf.ServiceException; @Category({RegionServerTests.class, MediumTests.class}) public class TestWALFiltering { diff --git a/pom.xml b/pom.xml index 0324c1c..0526e93 100644 --- a/pom.xml +++ b/pom.xml @@ -65,11 +65,15 @@ hbase-assembly hbase-testing-util hbase-annotations + hbase-checkstyle hbase-external-blockcache hbase-shaded + hbase-archetypes