From 8e3af510cab2eebd6bf2d08756df5fcfcba0fc03 Mon Sep 17 00:00:00 2001 From: Yi Liang Date: Mon, 17 Jul 2017 13:24:38 -0700 Subject: [PATCH] HBASE-18107: [AMv2] Remove DispatchMergingRegionsRequest & DispatchMergingRegions --- .../hbase/client/ConnectionImplementation.java | 7 - .../hbase/client/ShortCircuitMasterConnection.java | 6 - .../hadoop/hbase/shaded/protobuf/ProtobufUtil.java | 63 - .../shaded/protobuf/generated/AdminProtos.java | 1592 +------ .../shaded/protobuf/generated/MasterProtos.java | 4372 ++++++-------------- .../src/main/protobuf/Admin.proto | 19 - .../src/main/protobuf/Master.proto | 18 - .../org/apache/hadoop/hbase/master/HMaster.java | 54 - .../hadoop/hbase/master/MasterCoprocessorHost.java | 22 - .../hadoop/hbase/master/MasterRpcServices.java | 30 - .../apache/hadoop/hbase/master/MasterServices.java | 17 - .../apache/hadoop/hbase/master/ServerManager.java | 31 - .../procedure/DispatchMergingRegionsProcedure.java | 584 --- .../hadoop/hbase/regionserver/CompactSplit.java | 14 - .../hadoop/hbase/regionserver/RSRpcServices.java | 44 - .../hbase/regionserver/RegionMergeRequest.java | 108 - .../hbase/master/MockNoopMasterServices.java | 6 - .../hadoop/hbase/master/MockRegionServer.java | 9 - .../replication/regionserver/TestReplicator.java | 7 - 19 files changed, 1404 insertions(+), 5599 deletions(-) delete mode 100644 hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java delete mode 100644 hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java index 44974cb..4e50ae7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java @@ -1342,13 +1342,6 @@ class ConnectionImplementation implements ClusterConnection, Closeable { } @Override - public MasterProtos.DispatchMergingRegionsResponse dispatchMergingRegions( - RpcController controller, MasterProtos.DispatchMergingRegionsRequest request) - throws ServiceException { - return stub.dispatchMergingRegions(controller, request); - } - - @Override public MasterProtos.AssignRegionResponse assignRegion(RpcController controller, MasterProtos.AssignRegionRequest request) throws ServiceException { return stub.assignRegion(controller, request); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.java index 6d75446..a8050d4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.java @@ -505,10 +505,4 @@ public class ShortCircuitMasterConnection implements MasterKeepAliveConnection { throws ServiceException { return stub.splitRegion(controller, request); } - - @Override - public DispatchMergingRegionsResponse dispatchMergingRegions(RpcController controller, - DispatchMergingRegionsRequest request) throws ServiceException { - return stub.dispatchMergingRegions(controller, request); - } } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java index b1b52b1..91b7d3f 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java @@ -95,7 +95,6 @@ import org.apache.hadoop.hbase.io.TimeRange; import org.apache.hadoop.hbase.master.RegionState; import org.apache.hadoop.hbase.procedure2.LockInfo; import org.apache.hadoop.hbase.protobuf.ProtobufMagic; -import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest; import org.apache.hadoop.hbase.quotas.QuotaScope; import org.apache.hadoop.hbase.quotas.QuotaType; import org.apache.hadoop.hbase.quotas.SpaceViolationPolicy; @@ -2024,46 +2023,6 @@ public final class ProtobufUtil { } } - /** - * A helper to merge regions using admin protocol. Send request to - * regionserver. - * @param admin - * @param region_a - * @param region_b - * @param forcible true if do a compulsory merge, otherwise we will only merge - * two adjacent regions - * @param user effective user - * @throws IOException - */ - public static void mergeRegions(final RpcController controller, - final AdminService.BlockingInterface admin, - final HRegionInfo region_a, final HRegionInfo region_b, - final boolean forcible, final User user) throws IOException { - final MergeRegionsRequest request = ProtobufUtil.buildMergeRegionsRequest( - region_a.getRegionName(), region_b.getRegionName(),forcible); - if (user != null) { - try { - user.runAs(new PrivilegedExceptionAction() { - @Override - public Void run() throws Exception { - admin.mergeRegions(controller, request); - return null; - } - }); - } catch (InterruptedException ie) { - InterruptedIOException iioe = new InterruptedIOException(); - iioe.initCause(ie); - throw iioe; - } - } else { - try { - admin.mergeRegions(controller, request); - } catch (ServiceException se) { - throw ProtobufUtil.getRemoteException(se); - } - } - } - // End helpers for Admin /* @@ -3393,28 +3352,6 @@ public final class ProtobufUtil { return builder.build(); } - /** - * Create a MergeRegionsRequest for the given regions - * @param regionA name of region a - * @param regionB name of region b - * @param forcible true if it is a compulsory merge - * @return a MergeRegionsRequest - */ - public static MergeRegionsRequest buildMergeRegionsRequest( - final byte[] regionA, final byte[] regionB, final boolean forcible) { - MergeRegionsRequest.Builder builder = MergeRegionsRequest.newBuilder(); - RegionSpecifier regionASpecifier = RequestConverter.buildRegionSpecifier( - RegionSpecifierType.REGION_NAME, regionA); - RegionSpecifier regionBSpecifier = RequestConverter.buildRegionSpecifier( - RegionSpecifierType.REGION_NAME, regionB); - builder.setRegionA(regionASpecifier); - builder.setRegionB(regionBSpecifier); - builder.setForcible(forcible); - // send the master's wall clock time as well, so that the RS can refer to it - builder.setMasterSystemTime(EnvironmentEdgeManager.currentTime()); - return builder.build(); - } - /** * Get a ServerName from the passed in data bytes. * @param data Data with a serialize server name in it; can handle the old style diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java index 639be5a..5674591 100644 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java +++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java @@ -27150,1406 +27150,6 @@ public final class AdminProtos { } - public interface MergeRegionsRequestOrBuilder extends - // @@protoc_insertion_point(interface_extends:hbase.pb.MergeRegionsRequest) - org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { - - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - boolean hasRegionA(); - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); - - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - boolean hasRegionB(); - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); - - /** - * optional bool forcible = 3 [default = false]; - */ - boolean hasForcible(); - /** - * optional bool forcible = 3 [default = false]; - */ - boolean getForcible(); - - /** - *
-     * wall clock time from master
-     * 
- * - * optional uint64 master_system_time = 4; - */ - boolean hasMasterSystemTime(); - /** - *
-     * wall clock time from master
-     * 
- * - * optional uint64 master_system_time = 4; - */ - long getMasterSystemTime(); - } - /** - *
-   **
-   * Merges the specified regions.
-   * <p>
-   * This method currently closes the regions and then merges them
-   * 
- * - * Protobuf type {@code hbase.pb.MergeRegionsRequest} - */ - public static final class MergeRegionsRequest extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:hbase.pb.MergeRegionsRequest) - MergeRegionsRequestOrBuilder { - // Use MergeRegionsRequest.newBuilder() to construct. - private MergeRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { - super(builder); - } - private MergeRegionsRequest() { - forcible_ = false; - masterSystemTime_ = 0L; - } - - @java.lang.Override - public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private MergeRegionsRequest( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - this(); - int mutable_bitField0_ = 0; - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - subBuilder = regionA_.toBuilder(); - } - regionA_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(regionA_); - regionA_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000001; - break; - } - case 18: { - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - subBuilder = regionB_.toBuilder(); - } - regionB_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(regionB_); - regionB_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000002; - break; - } - case 24: { - bitField0_ |= 0x00000004; - forcible_ = input.readBool(); - break; - } - case 32: { - bitField0_ |= 0x00000008; - masterSystemTime_ = input.readUInt64(); - break; - } - } - } - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( - e).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.Builder.class); - } - - private int bitField0_; - public static final int REGION_A_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public boolean hasRegionA() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() { - return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() { - return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; - } - - public static final int REGION_B_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public boolean hasRegionB() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() { - return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() { - return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; - } - - public static final int FORCIBLE_FIELD_NUMBER = 3; - private boolean forcible_; - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean hasForcible() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean getForcible() { - return forcible_; - } - - public static final int MASTER_SYSTEM_TIME_FIELD_NUMBER = 4; - private long masterSystemTime_; - /** - *
-     * wall clock time from master
-     * 
- * - * optional uint64 master_system_time = 4; - */ - public boolean hasMasterSystemTime() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - *
-     * wall clock time from master
-     * 
- * - * optional uint64 master_system_time = 4; - */ - public long getMasterSystemTime() { - return masterSystemTime_; - } - - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized == 1) return true; - if (isInitialized == 0) return false; - - if (!hasRegionA()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasRegionB()) { - memoizedIsInitialized = 0; - return false; - } - if (!getRegionA().isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - if (!getRegionB().isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, getRegionA()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, getRegionB()); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBool(3, forcible_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeUInt64(4, masterSystemTime_); - } - unknownFields.writeTo(output); - } - - public int getSerializedSize() { - int size = memoizedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeMessageSize(1, getRegionA()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeMessageSize(2, getRegionB()); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeBoolSize(3, forcible_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeUInt64Size(4, masterSystemTime_); - } - size += unknownFields.getSerializedSize(); - memoizedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest) obj; - - boolean result = true; - result = result && (hasRegionA() == other.hasRegionA()); - if (hasRegionA()) { - result = result && getRegionA() - .equals(other.getRegionA()); - } - result = result && (hasRegionB() == other.hasRegionB()); - if (hasRegionB()) { - result = result && getRegionB() - .equals(other.getRegionB()); - } - result = result && (hasForcible() == other.hasForcible()); - if (hasForcible()) { - result = result && (getForcible() - == other.getForcible()); - } - result = result && (hasMasterSystemTime() == other.hasMasterSystemTime()); - if (hasMasterSystemTime()) { - result = result && (getMasterSystemTime() - == other.getMasterSystemTime()); - } - result = result && unknownFields.equals(other.unknownFields); - return result; - } - - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptor().hashCode(); - if (hasRegionA()) { - hash = (37 * hash) + REGION_A_FIELD_NUMBER; - hash = (53 * hash) + getRegionA().hashCode(); - } - if (hasRegionB()) { - hash = (37 * hash) + REGION_B_FIELD_NUMBER; - hash = (53 * hash) + getRegionB().hashCode(); - } - if (hasForcible()) { - hash = (37 * hash) + FORCIBLE_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean( - getForcible()); - } - if (hasMasterSystemTime()) { - hash = (37 * hash) + MASTER_SYSTEM_TIME_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong( - getMasterSystemTime()); - } - hash = (29 * hash) + unknownFields.hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(byte[] data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom( - byte[] data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseDelimitedFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder() { - return DEFAULT_INSTANCE.toBuilder(); - } - public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest prototype) { - return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { - return this == DEFAULT_INSTANCE - ? new Builder() : new Builder().mergeFrom(this); - } - - @java.lang.Override - protected Builder newBuilderForType( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - *
-     **
-     * Merges the specified regions.
-     * <p>
-     * This method currently closes the regions and then merges them
-     * 
- * - * Protobuf type {@code hbase.pb.MergeRegionsRequest} - */ - public static final class Builder extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder implements - // @@protoc_insertion_point(builder_implements:hbase.pb.MergeRegionsRequest) - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequestOrBuilder { - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .alwaysUseFieldBuilders) { - getRegionAFieldBuilder(); - getRegionBFieldBuilder(); - } - } - public Builder clear() { - super.clear(); - if (regionABuilder_ == null) { - regionA_ = null; - } else { - regionABuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000001); - if (regionBBuilder_ == null) { - regionB_ = null; - } else { - regionBBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - forcible_ = false; - bitField0_ = (bitField0_ & ~0x00000004); - masterSystemTime_ = 0L; - bitField0_ = (bitField0_ & ~0x00000008); - return this; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsRequest_descriptor; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest build() { - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest buildPartial() { - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - if (regionABuilder_ == null) { - result.regionA_ = regionA_; - } else { - result.regionA_ = regionABuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - if (regionBBuilder_ == null) { - result.regionB_ = regionB_; - } else { - result.regionB_ = regionBBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.forcible_ = forcible_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; - } - result.masterSystemTime_ = masterSystemTime_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder clone() { - return (Builder) super.clone(); - } - public Builder setField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.setField(field, value); - } - public Builder clearField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) { - return (Builder) super.clearField(field); - } - public Builder clearOneof( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) { - return (Builder) super.clearOneof(oneof); - } - public Builder setRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - int index, Object value) { - return (Builder) super.setRepeatedField(field, index, value); - } - public Builder addRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.addRepeatedField(field, value); - } - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest) { - return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest other) { - if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance()) return this; - if (other.hasRegionA()) { - mergeRegionA(other.getRegionA()); - } - if (other.hasRegionB()) { - mergeRegionB(other.getRegionB()); - } - if (other.hasForcible()) { - setForcible(other.getForcible()); - } - if (other.hasMasterSystemTime()) { - setMasterSystemTime(other.getMasterSystemTime()); - } - this.mergeUnknownFields(other.unknownFields); - onChanged(); - return this; - } - - public final boolean isInitialized() { - if (!hasRegionA()) { - return false; - } - if (!hasRegionB()) { - return false; - } - if (!getRegionA().isInitialized()) { - return false; - } - if (!getRegionB().isInitialized()) { - return false; - } - return true; - } - - public Builder mergeFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest) e.getUnfinishedMessage(); - throw e.unwrapIOException(); - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = null; - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public boolean hasRegionA() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() { - if (regionABuilder_ == null) { - return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; - } else { - return regionABuilder_.getMessage(); - } - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public Builder setRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionABuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - regionA_ = value; - onChanged(); - } else { - regionABuilder_.setMessage(value); - } - bitField0_ |= 0x00000001; - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public Builder setRegionA( - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionABuilder_ == null) { - regionA_ = builderForValue.build(); - onChanged(); - } else { - regionABuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000001; - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public Builder mergeRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionABuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && - regionA_ != null && - regionA_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { - regionA_ = - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionA_).mergeFrom(value).buildPartial(); - } else { - regionA_ = value; - } - onChanged(); - } else { - regionABuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000001; - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public Builder clearRegionA() { - if (regionABuilder_ == null) { - regionA_ = null; - onChanged(); - } else { - regionABuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000001); - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionABuilder() { - bitField0_ |= 0x00000001; - onChanged(); - return getRegionAFieldBuilder().getBuilder(); - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() { - if (regionABuilder_ != null) { - return regionABuilder_.getMessageOrBuilder(); - } else { - return regionA_ == null ? - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; - } - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> - getRegionAFieldBuilder() { - if (regionABuilder_ == null) { - regionABuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( - getRegionA(), - getParentForChildren(), - isClean()); - regionA_ = null; - } - return regionABuilder_; - } - - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = null; - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public boolean hasRegionB() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() { - if (regionBBuilder_ == null) { - return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; - } else { - return regionBBuilder_.getMessage(); - } - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public Builder setRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - regionB_ = value; - onChanged(); - } else { - regionBBuilder_.setMessage(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public Builder setRegionB( - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionBBuilder_ == null) { - regionB_ = builderForValue.build(); - onChanged(); - } else { - regionBBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public Builder mergeRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && - regionB_ != null && - regionB_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { - regionB_ = - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionB_).mergeFrom(value).buildPartial(); - } else { - regionB_ = value; - } - onChanged(); - } else { - regionBBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public Builder clearRegionB() { - if (regionBBuilder_ == null) { - regionB_ = null; - onChanged(); - } else { - regionBBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBBuilder() { - bitField0_ |= 0x00000002; - onChanged(); - return getRegionBFieldBuilder().getBuilder(); - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() { - if (regionBBuilder_ != null) { - return regionBBuilder_.getMessageOrBuilder(); - } else { - return regionB_ == null ? - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; - } - } - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> - getRegionBFieldBuilder() { - if (regionBBuilder_ == null) { - regionBBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( - getRegionB(), - getParentForChildren(), - isClean()); - regionB_ = null; - } - return regionBBuilder_; - } - - private boolean forcible_ ; - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean hasForcible() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean getForcible() { - return forcible_; - } - /** - * optional bool forcible = 3 [default = false]; - */ - public Builder setForcible(boolean value) { - bitField0_ |= 0x00000004; - forcible_ = value; - onChanged(); - return this; - } - /** - * optional bool forcible = 3 [default = false]; - */ - public Builder clearForcible() { - bitField0_ = (bitField0_ & ~0x00000004); - forcible_ = false; - onChanged(); - return this; - } - - private long masterSystemTime_ ; - /** - *
-       * wall clock time from master
-       * 
- * - * optional uint64 master_system_time = 4; - */ - public boolean hasMasterSystemTime() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - *
-       * wall clock time from master
-       * 
- * - * optional uint64 master_system_time = 4; - */ - public long getMasterSystemTime() { - return masterSystemTime_; - } - /** - *
-       * wall clock time from master
-       * 
- * - * optional uint64 master_system_time = 4; - */ - public Builder setMasterSystemTime(long value) { - bitField0_ |= 0x00000008; - masterSystemTime_ = value; - onChanged(); - return this; - } - /** - *
-       * wall clock time from master
-       * 
- * - * optional uint64 master_system_time = 4; - */ - public Builder clearMasterSystemTime() { - bitField0_ = (bitField0_ & ~0x00000008); - masterSystemTime_ = 0L; - onChanged(); - return this; - } - public final Builder setUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.setUnknownFields(unknownFields); - } - - public final Builder mergeUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.mergeUnknownFields(unknownFields); - } - - - // @@protoc_insertion_point(builder_scope:hbase.pb.MergeRegionsRequest) - } - - // @@protoc_insertion_point(class_scope:hbase.pb.MergeRegionsRequest) - private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest DEFAULT_INSTANCE; - static { - DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest(); - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest getDefaultInstance() { - return DEFAULT_INSTANCE; - } - - @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser - PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser() { - public MergeRegionsRequest parsePartialFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return new MergeRegionsRequest(input, extensionRegistry); - } - }; - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser parser() { - return PARSER; - } - - @java.lang.Override - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest getDefaultInstanceForType() { - return DEFAULT_INSTANCE; - } - - } - - public interface MergeRegionsResponseOrBuilder extends - // @@protoc_insertion_point(interface_extends:hbase.pb.MergeRegionsResponse) - org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { - } - /** - * Protobuf type {@code hbase.pb.MergeRegionsResponse} - */ - public static final class MergeRegionsResponse extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:hbase.pb.MergeRegionsResponse) - MergeRegionsResponseOrBuilder { - // Use MergeRegionsResponse.newBuilder() to construct. - private MergeRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { - super(builder); - } - private MergeRegionsResponse() { - } - - @java.lang.Override - public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private MergeRegionsResponse( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - this(); - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - } - } - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( - e).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.Builder.class); - } - - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized == 1) return true; - if (isInitialized == 0) return false; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - unknownFields.writeTo(output); - } - - public int getSerializedSize() { - int size = memoizedSize; - if (size != -1) return size; - - size = 0; - size += unknownFields.getSerializedSize(); - memoizedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) obj; - - boolean result = true; - result = result && unknownFields.equals(other.unknownFields); - return result; - } - - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptor().hashCode(); - hash = (29 * hash) + unknownFields.hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(byte[] data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom( - byte[] data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseDelimitedFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder() { - return DEFAULT_INSTANCE.toBuilder(); - } - public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse prototype) { - return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { - return this == DEFAULT_INSTANCE - ? new Builder() : new Builder().mergeFrom(this); - } - - @java.lang.Override - protected Builder newBuilderForType( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.MergeRegionsResponse} - */ - public static final class Builder extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder implements - // @@protoc_insertion_point(builder_implements:hbase.pb.MergeRegionsResponse) - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponseOrBuilder { - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .alwaysUseFieldBuilders) { - } - } - public Builder clear() { - super.clear(); - return this; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_MergeRegionsResponse_descriptor; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse getDefaultInstanceForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse build() { - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse buildPartial() { - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse(this); - onBuilt(); - return result; - } - - public Builder clone() { - return (Builder) super.clone(); - } - public Builder setField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.setField(field, value); - } - public Builder clearField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) { - return (Builder) super.clearField(field); - } - public Builder clearOneof( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) { - return (Builder) super.clearOneof(oneof); - } - public Builder setRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - int index, Object value) { - return (Builder) super.setRepeatedField(field, index, value); - } - public Builder addRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.addRepeatedField(field, value); - } - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) { - return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse other) { - if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance()) return this; - this.mergeUnknownFields(other.unknownFields); - onChanged(); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) e.getUnfinishedMessage(); - throw e.unwrapIOException(); - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - public final Builder setUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.setUnknownFields(unknownFields); - } - - public final Builder mergeUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.mergeUnknownFields(unknownFields); - } - - - // @@protoc_insertion_point(builder_scope:hbase.pb.MergeRegionsResponse) - } - - // @@protoc_insertion_point(class_scope:hbase.pb.MergeRegionsResponse) - private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse DEFAULT_INSTANCE; - static { - DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse(); - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse getDefaultInstance() { - return DEFAULT_INSTANCE; - } - - @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser - PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser() { - public MergeRegionsResponse parsePartialFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return new MergeRegionsResponse(input, extensionRegistry); - } - }; - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser parser() { - return PARSER; - } - - @java.lang.Override - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse getDefaultInstanceForType() { - return DEFAULT_INSTANCE; - } - - } - /** * Protobuf service {@code hbase.pb.AdminService} */ @@ -28722,14 +27322,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done); - /** - * rpc MergeRegions(.hbase.pb.MergeRegionsRequest) returns (.hbase.pb.MergeRegionsResponse); - */ - public abstract void mergeRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done); - } public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Service newReflectiveService( @@ -28895,14 +27487,6 @@ public final class AdminProtos { impl.executeProcedures(controller, request, done); } - @java.lang.Override - public void mergeRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { - impl.mergeRegions(controller, request, done); - } - }; } @@ -28965,8 +27549,6 @@ public final class AdminProtos { return impl.getSpaceQuotaSnapshots(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsRequest)request); case 19: return impl.executeProcedures(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest)request); - case 20: - return impl.mergeRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)request); default: throw new java.lang.AssertionError("Can't get here."); } @@ -29021,8 +27603,6 @@ public final class AdminProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsRequest.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest.getDefaultInstance(); - case 20: - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } @@ -29077,8 +27657,6 @@ public final class AdminProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsResponse.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse.getDefaultInstance(); - case 20: - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } @@ -29251,14 +27829,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done); - /** - * rpc MergeRegions(.hbase.pb.MergeRegionsRequest) returns (.hbase.pb.MergeRegionsResponse); - */ - public abstract void mergeRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done); - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { @@ -29381,11 +27951,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 20: - this.mergeRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest)request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( - done)); - return; default: throw new java.lang.AssertionError("Can't get here."); } @@ -29440,8 +28005,6 @@ public final class AdminProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsRequest.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest.getDefaultInstance(); - case 20: - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } @@ -29496,8 +28059,6 @@ public final class AdminProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsResponse.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse.getDefaultInstance(); - case 20: - return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } @@ -29818,21 +28379,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse.getDefaultInstance())); } - - public void mergeRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { - channel.callMethod( - getDescriptor().getMethods().get(20), - controller, - request, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance(), - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( - done, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.class, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance())); - } } public static BlockingInterface newBlockingStub( @@ -29940,11 +28486,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse mergeRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { @@ -30193,18 +28734,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse.getDefaultInstance()); } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse mergeRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest request) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { - return (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(20), - controller, - request, - org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse.getDefaultInstance()); - } - } // @@protoc_insertion_point(class_scope:hbase.pb.AdminService) @@ -30410,16 +28939,6 @@ public final class AdminProtos { private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ExecuteProceduresResponse_fieldAccessorTable; - private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_MergeRegionsRequest_descriptor; - private static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_hbase_pb_MergeRegionsRequest_fieldAccessorTable; - private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_MergeRegionsResponse_descriptor; - private static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_hbase_pb_MergeRegionsResponse_fieldAccessorTable; public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { @@ -30513,59 +29032,52 @@ public final class AdminProtos { "nRequest\"\203\001\n\031ExecuteProceduresResponse\0221" + "\n\013open_region\030\001 \003(\0132\034.hbase.pb.OpenRegio" + "nResponse\0223\n\014close_region\030\002 \003(\0132\035.hbase." + - "pb.CloseRegionResponse\"\244\001\n\023MergeRegionsR" + - "equest\022+\n\010region_a\030\001 \002(\0132\031.hbase.pb.Regi" + - "onSpecifier\022+\n\010region_b\030\002 \002(\0132\031.hbase.pb" + - ".RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005fals" + - "e\022\032\n\022master_system_time\030\004 \001(\004\"\026\n\024MergeRe" + - "gionsResponse2\216\016\n\014AdminService\022P\n\rGetReg", - "ionInfo\022\036.hbase.pb.GetRegionInfoRequest\032" + - "\037.hbase.pb.GetRegionInfoResponse\022M\n\014GetS" + - "toreFile\022\035.hbase.pb.GetStoreFileRequest\032" + - "\036.hbase.pb.GetStoreFileResponse\022V\n\017GetOn" + - "lineRegion\022 .hbase.pb.GetOnlineRegionReq" + - "uest\032!.hbase.pb.GetOnlineRegionResponse\022" + - "G\n\nOpenRegion\022\033.hbase.pb.OpenRegionReque" + - "st\032\034.hbase.pb.OpenRegionResponse\022M\n\014Warm" + - "upRegion\022\035.hbase.pb.WarmupRegionRequest\032" + - "\036.hbase.pb.WarmupRegionResponse\022J\n\013Close", - "Region\022\034.hbase.pb.CloseRegionRequest\032\035.h" + - "base.pb.CloseRegionResponse\022J\n\013FlushRegi" + - "on\022\034.hbase.pb.FlushRegionRequest\032\035.hbase" + - ".pb.FlushRegionResponse\022J\n\013SplitRegion\022\034" + - ".hbase.pb.SplitRegionRequest\032\035.hbase.pb." + - "SplitRegionResponse\022P\n\rCompactRegion\022\036.h" + - "base.pb.CompactRegionRequest\032\037.hbase.pb." + - "CompactRegionResponse\022\\\n\021ReplicateWALEnt" + - "ry\022\".hbase.pb.ReplicateWALEntryRequest\032#" + - ".hbase.pb.ReplicateWALEntryResponse\022Q\n\006R", - "eplay\022\".hbase.pb.ReplicateWALEntryReques" + - "t\032#.hbase.pb.ReplicateWALEntryResponse\022P" + - "\n\rRollWALWriter\022\036.hbase.pb.RollWALWriter" + - "Request\032\037.hbase.pb.RollWALWriterResponse" + - "\022P\n\rGetServerInfo\022\036.hbase.pb.GetServerIn" + - "foRequest\032\037.hbase.pb.GetServerInfoRespon" + - "se\022G\n\nStopServer\022\033.hbase.pb.StopServerRe" + - "quest\032\034.hbase.pb.StopServerResponse\022_\n\022U" + - "pdateFavoredNodes\022#.hbase.pb.UpdateFavor" + - "edNodesRequest\032$.hbase.pb.UpdateFavoredN", - "odesResponse\022b\n\023UpdateConfiguration\022$.hb" + - "ase.pb.UpdateConfigurationRequest\032%.hbas" + - "e.pb.UpdateConfigurationResponse\022P\n\rGetR" + - "egionLoad\022\036.hbase.pb.GetRegionLoadReques" + - "t\032\037.hbase.pb.GetRegionLoadResponse\022h\n\025Cl" + - "earCompactionQueues\022&.hbase.pb.ClearComp" + - "actionQueuesRequest\032\'.hbase.pb.ClearComp" + - "actionQueuesResponse\022k\n\026GetSpaceQuotaSna" + - "pshots\022\'.hbase.pb.GetSpaceQuotaSnapshots" + - "Request\032(.hbase.pb.GetSpaceQuotaSnapshot", - "sResponse\022\\\n\021ExecuteProcedures\022\".hbase.p" + - "b.ExecuteProceduresRequest\032#.hbase.pb.Ex" + - "ecuteProceduresResponse\022M\n\014MergeRegions\022" + - "\035.hbase.pb.MergeRegionsRequest\032\036.hbase.p" + - "b.MergeRegionsResponseBH\n1org.apache.had" + - "oop.hbase.shaded.protobuf.generatedB\013Adm" + - "inProtosH\001\210\001\001\240\001\001" + "pb.CloseRegionResponse2\277\r\n\014AdminService\022" + + "P\n\rGetRegionInfo\022\036.hbase.pb.GetRegionInf" + + "oRequest\032\037.hbase.pb.GetRegionInfoRespons" + + "e\022M\n\014GetStoreFile\022\035.hbase.pb.GetStoreFil" + + "eRequest\032\036.hbase.pb.GetStoreFileResponse" + + "\022V\n\017GetOnlineRegion\022 .hbase.pb.GetOnline", + "RegionRequest\032!.hbase.pb.GetOnlineRegion" + + "Response\022G\n\nOpenRegion\022\033.hbase.pb.OpenRe" + + "gionRequest\032\034.hbase.pb.OpenRegionRespons" + + "e\022M\n\014WarmupRegion\022\035.hbase.pb.WarmupRegio" + + "nRequest\032\036.hbase.pb.WarmupRegionResponse" + + "\022J\n\013CloseRegion\022\034.hbase.pb.CloseRegionRe" + + "quest\032\035.hbase.pb.CloseRegionResponse\022J\n\013" + + "FlushRegion\022\034.hbase.pb.FlushRegionReques" + + "t\032\035.hbase.pb.FlushRegionResponse\022J\n\013Spli" + + "tRegion\022\034.hbase.pb.SplitRegionRequest\032\035.", + "hbase.pb.SplitRegionResponse\022P\n\rCompactR" + + "egion\022\036.hbase.pb.CompactRegionRequest\032\037." + + "hbase.pb.CompactRegionResponse\022\\\n\021Replic" + + "ateWALEntry\022\".hbase.pb.ReplicateWALEntry" + + "Request\032#.hbase.pb.ReplicateWALEntryResp" + + "onse\022Q\n\006Replay\022\".hbase.pb.ReplicateWALEn" + + "tryRequest\032#.hbase.pb.ReplicateWALEntryR" + + "esponse\022P\n\rRollWALWriter\022\036.hbase.pb.Roll" + + "WALWriterRequest\032\037.hbase.pb.RollWALWrite" + + "rResponse\022P\n\rGetServerInfo\022\036.hbase.pb.Ge", + "tServerInfoRequest\032\037.hbase.pb.GetServerI" + + "nfoResponse\022G\n\nStopServer\022\033.hbase.pb.Sto" + + "pServerRequest\032\034.hbase.pb.StopServerResp" + + "onse\022_\n\022UpdateFavoredNodes\022#.hbase.pb.Up" + + "dateFavoredNodesRequest\032$.hbase.pb.Updat" + + "eFavoredNodesResponse\022b\n\023UpdateConfigura" + + "tion\022$.hbase.pb.UpdateConfigurationReque" + + "st\032%.hbase.pb.UpdateConfigurationRespons" + + "e\022P\n\rGetRegionLoad\022\036.hbase.pb.GetRegionL" + + "oadRequest\032\037.hbase.pb.GetRegionLoadRespo", + "nse\022h\n\025ClearCompactionQueues\022&.hbase.pb." + + "ClearCompactionQueuesRequest\032\'.hbase.pb." + + "ClearCompactionQueuesResponse\022k\n\026GetSpac" + + "eQuotaSnapshots\022\'.hbase.pb.GetSpaceQuota" + + "SnapshotsRequest\032(.hbase.pb.GetSpaceQuot" + + "aSnapshotsResponse\022\\\n\021ExecuteProcedures\022" + + "\".hbase.pb.ExecuteProceduresRequest\032#.hb" + + "ase.pb.ExecuteProceduresResponseBH\n1org." + + "apache.hadoop.hbase.shaded.protobuf.gene" + + "ratedB\013AdminProtosH\001\210\001\001\240\001\001" }; org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() { @@ -30823,18 +29335,6 @@ public final class AdminProtos { org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ExecuteProceduresResponse_descriptor, new java.lang.String[] { "OpenRegion", "CloseRegion", }); - internal_static_hbase_pb_MergeRegionsRequest_descriptor = - getDescriptor().getMessageTypes().get(38); - internal_static_hbase_pb_MergeRegionsRequest_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_hbase_pb_MergeRegionsRequest_descriptor, - new java.lang.String[] { "RegionA", "RegionB", "Forcible", "MasterSystemTime", }); - internal_static_hbase_pb_MergeRegionsResponse_descriptor = - getDescriptor().getMessageTypes().get(39); - internal_static_hbase_pb_MergeRegionsResponse_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_hbase_pb_MergeRegionsResponse_descriptor, - new java.lang.String[] { }); org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.getDescriptor(); diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java index 0d2e09f..bc401c2 100644 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java +++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java @@ -5507,35 +5507,33 @@ public final class MasterProtos { } - public interface DispatchMergingRegionsRequestOrBuilder extends - // @@protoc_insertion_point(interface_extends:hbase.pb.DispatchMergingRegionsRequest) + public interface MergeTableRegionsRequestOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsRequest) org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - boolean hasRegionA(); - /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); + java.util.List + getRegionList(); /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); - + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index); /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - boolean hasRegionB(); + int getRegionCount(); /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); + java.util.List + getRegionOrBuilderList(); /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( + int index); /** * optional bool forcible = 3 [default = false]; @@ -5567,20 +5565,21 @@ public final class MasterProtos { /** *
    **
-   * Dispatch merging the specified regions.
+   * Merging the specified regions in a table.
    * 
* - * Protobuf type {@code hbase.pb.DispatchMergingRegionsRequest} + * Protobuf type {@code hbase.pb.MergeTableRegionsRequest} */ - public static final class DispatchMergingRegionsRequest extends + public static final class MergeTableRegionsRequest extends org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:hbase.pb.DispatchMergingRegionsRequest) - DispatchMergingRegionsRequestOrBuilder { - // Use DispatchMergingRegionsRequest.newBuilder() to construct. - private DispatchMergingRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { + // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsRequest) + MergeTableRegionsRequestOrBuilder { + // Use MergeTableRegionsRequest.newBuilder() to construct. + private MergeTableRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } - private DispatchMergingRegionsRequest() { + private MergeTableRegionsRequest() { + region_ = java.util.Collections.emptyList(); forcible_ = false; nonceGroup_ = 0L; nonce_ = 0L; @@ -5591,7 +5590,7 @@ public final class MasterProtos { getUnknownFields() { return this.unknownFields; } - private DispatchMergingRegionsRequest( + private MergeTableRegionsRequest( org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { @@ -5615,43 +5614,26 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - subBuilder = regionA_.toBuilder(); - } - regionA_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(regionA_); - regionA_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000001; - break; - } - case 18: { - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - subBuilder = regionB_.toBuilder(); - } - regionB_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(regionB_); - regionB_ = subBuilder.buildPartial(); + if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + region_ = new java.util.ArrayList(); + mutable_bitField0_ |= 0x00000001; } - bitField0_ |= 0x00000002; + region_.add( + input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry)); break; } case 24: { - bitField0_ |= 0x00000004; + bitField0_ |= 0x00000001; forcible_ = input.readBool(); break; } case 32: { - bitField0_ |= 0x00000008; + bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 40: { - bitField0_ |= 0x00000010; + bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } @@ -5663,63 +5645,59 @@ public final class MasterProtos { throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { + if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + region_ = java.util.Collections.unmodifiableList(region_); + } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor; + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; } protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_fieldAccessorTable + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.Builder.class); + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class); } private int bitField0_; - public static final int REGION_A_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; - /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public boolean hasRegionA() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } + public static final int REGION_FIELD_NUMBER = 1; + private java.util.List region_; /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() { - return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; + public java.util.List getRegionList() { + return region_; } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() { - return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; + public java.util.List + getRegionOrBuilderList() { + return region_; } - - public static final int REGION_B_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public boolean hasRegionB() { - return ((bitField0_ & 0x00000002) == 0x00000002); + public int getRegionCount() { + return region_.size(); } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() { - return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { + return region_.get(index); } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() { - return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( + int index) { + return region_.get(index); } public static final int FORCIBLE_FIELD_NUMBER = 3; @@ -5728,7 +5706,7 @@ public final class MasterProtos { * optional bool forcible = 3 [default = false]; */ public boolean hasForcible() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000001) == 0x00000001); } /** * optional bool forcible = 3 [default = false]; @@ -5743,7 +5721,7 @@ public final class MasterProtos { * optional uint64 nonce_group = 4 [default = 0]; */ public boolean hasNonceGroup() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000002) == 0x00000002); } /** * optional uint64 nonce_group = 4 [default = 0]; @@ -5758,7 +5736,7 @@ public final class MasterProtos { * optional uint64 nonce = 5 [default = 0]; */ public boolean hasNonce() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000004) == 0x00000004); } /** * optional uint64 nonce = 5 [default = 0]; @@ -5773,21 +5751,11 @@ public final class MasterProtos { if (isInitialized == 1) return true; if (isInitialized == 0) return false; - if (!hasRegionA()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasRegionB()) { - memoizedIsInitialized = 0; - return false; - } - if (!getRegionA().isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - if (!getRegionB().isInitialized()) { - memoizedIsInitialized = 0; - return false; + for (int i = 0; i < getRegionCount(); i++) { + if (!getRegion(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } } memoizedIsInitialized = 1; return true; @@ -5795,19 +5763,16 @@ public final class MasterProtos { public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, getRegionA()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, getRegionB()); + for (int i = 0; i < region_.size(); i++) { + output.writeMessage(1, region_.get(i)); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBool(3, forcible_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeUInt64(4, nonceGroup_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeUInt64(5, nonce_); } unknownFields.writeTo(output); @@ -5818,23 +5783,19 @@ public final class MasterProtos { if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeMessageSize(1, getRegionA()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + for (int i = 0; i < region_.size(); i++) { size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeMessageSize(2, getRegionB()); + .computeMessageSize(1, region_.get(i)); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000001) == 0x00000001)) { size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream .computeBoolSize(3, forcible_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000002) == 0x00000002)) { size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream .computeUInt64Size(4, nonceGroup_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + if (((bitField0_ & 0x00000004) == 0x00000004)) { size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream .computeUInt64Size(5, nonce_); } @@ -5849,22 +5810,14 @@ public final class MasterProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)) { + if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)) { return super.equals(obj); } - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) obj; + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) obj; boolean result = true; - result = result && (hasRegionA() == other.hasRegionA()); - if (hasRegionA()) { - result = result && getRegionA() - .equals(other.getRegionA()); - } - result = result && (hasRegionB() == other.hasRegionB()); - if (hasRegionB()) { - result = result && getRegionB() - .equals(other.getRegionB()); - } + result = result && getRegionList() + .equals(other.getRegionList()); result = result && (hasForcible() == other.hasForcible()); if (hasForcible()) { result = result && (getForcible() @@ -5891,13 +5844,9 @@ public final class MasterProtos { } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); - if (hasRegionA()) { - hash = (37 * hash) + REGION_A_FIELD_NUMBER; - hash = (53 * hash) + getRegionA().hashCode(); - } - if (hasRegionB()) { - hash = (37 * hash) + REGION_B_FIELD_NUMBER; - hash = (53 * hash) + getRegionB().hashCode(); + if (getRegionCount() > 0) { + hash = (37 * hash) + REGION_FIELD_NUMBER; + hash = (53 * hash) + getRegionList().hashCode(); } if (hasForcible()) { hash = (37 * hash) + FORCIBLE_FIELD_NUMBER; @@ -5919,58 +5868,58 @@ public final class MasterProtos { return hash; } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(byte[] data) + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(byte[] data) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( byte[] data, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( java.io.InputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseDelimitedFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -5982,7 +5931,7 @@ public final class MasterProtos { public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } public Builder toBuilder() { @@ -5999,28 +5948,28 @@ public final class MasterProtos { /** *
      **
-     * Dispatch merging the specified regions.
+     * Merging the specified regions in a table.
      * 
* - * Protobuf type {@code hbase.pb.DispatchMergingRegionsRequest} + * Protobuf type {@code hbase.pb.MergeTableRegionsRequest} */ public static final class Builder extends org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder implements - // @@protoc_insertion_point(builder_implements:hbase.pb.DispatchMergingRegionsRequest) - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequestOrBuilder { + // @@protoc_insertion_point(builder_implements:hbase.pb.MergeTableRegionsRequest) + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequestOrBuilder { public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor; + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; } protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_fieldAccessorTable + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.Builder.class); + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class); } - // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.newBuilder() + // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -6033,80 +5982,66 @@ public final class MasterProtos { private void maybeForceBuilderInitialization() { if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { - getRegionAFieldBuilder(); - getRegionBFieldBuilder(); + getRegionFieldBuilder(); } } public Builder clear() { super.clear(); - if (regionABuilder_ == null) { - regionA_ = null; - } else { - regionABuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000001); - if (regionBBuilder_ == null) { - regionB_ = null; + if (regionBuilder_ == null) { + region_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000001); } else { - regionBBuilder_.clear(); + regionBuilder_.clear(); } - bitField0_ = (bitField0_ & ~0x00000002); forcible_ = false; - bitField0_ = (bitField0_ & ~0x00000004); + bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; - bitField0_ = (bitField0_ & ~0x00000008); + bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; - bitField0_ = (bitField0_ & ~0x00000010); + bitField0_ = (bitField0_ & ~0x00000008); return this; } public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor; + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; } - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance(); + public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() { + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); } - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest build() { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest result = buildPartial(); + public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest build() { + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest buildPartial() { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest(this); + public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest buildPartial() { + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - if (regionABuilder_ == null) { - result.regionA_ = regionA_; + if (regionBuilder_ == null) { + if (((bitField0_ & 0x00000001) == 0x00000001)) { + region_ = java.util.Collections.unmodifiableList(region_); + bitField0_ = (bitField0_ & ~0x00000001); + } + result.region_ = region_; } else { - result.regionA_ = regionABuilder_.build(); + result.region_ = regionBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - if (regionBBuilder_ == null) { - result.regionB_ = regionB_; - } else { - result.regionB_ = regionBBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; + to_bitField0_ |= 0x00000001; } result.forcible_ = forcible_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000002; } result.nonceGroup_ = nonceGroup_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000004; } result.nonce_ = nonce_; result.bitField0_ = to_bitField0_; @@ -6141,21 +6076,41 @@ public final class MasterProtos { return (Builder) super.addRepeatedField(field, value); } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) { - return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)other); + if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) { + return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other) { - if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance()) return this; - if (other.hasRegionA()) { - mergeRegionA(other.getRegionA()); - } - if (other.hasRegionB()) { - mergeRegionB(other.getRegionB()); + public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other) { + if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance()) return this; + if (regionBuilder_ == null) { + if (!other.region_.isEmpty()) { + if (region_.isEmpty()) { + region_ = other.region_; + bitField0_ = (bitField0_ & ~0x00000001); + } else { + ensureRegionIsMutable(); + region_.addAll(other.region_); + } + onChanged(); + } + } else { + if (!other.region_.isEmpty()) { + if (regionBuilder_.isEmpty()) { + regionBuilder_.dispose(); + regionBuilder_ = null; + region_ = other.region_; + bitField0_ = (bitField0_ & ~0x00000001); + regionBuilder_ = + org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? + getRegionFieldBuilder() : null; + } else { + regionBuilder_.addAllMessages(other.region_); + } + } } if (other.hasForcible()) { setForcible(other.getForcible()); @@ -6172,17 +6127,10 @@ public final class MasterProtos { } public final boolean isInitialized() { - if (!hasRegionA()) { - return false; - } - if (!hasRegionB()) { - return false; - } - if (!getRegionA().isInitialized()) { - return false; - } - if (!getRegionB().isInitialized()) { - return false; + for (int i = 0; i < getRegionCount(); i++) { + if (!getRegion(i).isInitialized()) { + return false; + } } return true; } @@ -6191,11 +6139,11 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parsedMessage = null; + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { @@ -6206,240 +6154,244 @@ public final class MasterProtos { } private int bitField0_; - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = null; - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; + private java.util.List region_ = + java.util.Collections.emptyList(); + private void ensureRegionIsMutable() { + if (!((bitField0_ & 0x00000001) == 0x00000001)) { + region_ = new java.util.ArrayList(region_); + bitField0_ |= 0x00000001; + } + } + + private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; + /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public boolean hasRegionA() { - return ((bitField0_ & 0x00000001) == 0x00000001); + public java.util.List getRegionList() { + if (regionBuilder_ == null) { + return java.util.Collections.unmodifiableList(region_); + } else { + return regionBuilder_.getMessageList(); + } } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() { - if (regionABuilder_ == null) { - return regionA_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; + public int getRegionCount() { + if (regionBuilder_ == null) { + return region_.size(); } else { - return regionABuilder_.getMessage(); + return regionBuilder_.getCount(); } } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder setRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionABuilder_ == null) { + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { + if (regionBuilder_ == null) { + return region_.get(index); + } else { + return regionBuilder_.getMessage(index); + } + } + /** + * repeated .hbase.pb.RegionSpecifier region = 1; + */ + public Builder setRegion( + int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { + if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } - regionA_ = value; + ensureRegionIsMutable(); + region_.set(index, value); onChanged(); } else { - regionABuilder_.setMessage(value); + regionBuilder_.setMessage(index, value); } - bitField0_ |= 0x00000001; return this; } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder setRegionA( - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionABuilder_ == null) { - regionA_ = builderForValue.build(); + public Builder setRegion( + int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { + if (regionBuilder_ == null) { + ensureRegionIsMutable(); + region_.set(index, builderForValue.build()); onChanged(); } else { - regionABuilder_.setMessage(builderForValue.build()); + regionBuilder_.setMessage(index, builderForValue.build()); } - bitField0_ |= 0x00000001; return this; } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder mergeRegionA(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionABuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && - regionA_ != null && - regionA_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { - regionA_ = - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionA_).mergeFrom(value).buildPartial(); - } else { - regionA_ = value; + public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { + if (regionBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); } + ensureRegionIsMutable(); + region_.add(value); onChanged(); } else { - regionABuilder_.mergeFrom(value); + regionBuilder_.addMessage(value); } - bitField0_ |= 0x00000001; return this; } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder clearRegionA() { - if (regionABuilder_ == null) { - regionA_ = null; + public Builder addRegion( + int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { + if (regionBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureRegionIsMutable(); + region_.add(index, value); onChanged(); } else { - regionABuilder_.clear(); + regionBuilder_.addMessage(index, value); } - bitField0_ = (bitField0_ & ~0x00000001); return this; } /** - * required .hbase.pb.RegionSpecifier region_a = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionABuilder() { - bitField0_ |= 0x00000001; - onChanged(); - return getRegionAFieldBuilder().getBuilder(); - } - /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() { - if (regionABuilder_ != null) { - return regionABuilder_.getMessageOrBuilder(); + public Builder addRegion( + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { + if (regionBuilder_ == null) { + ensureRegionIsMutable(); + region_.add(builderForValue.build()); + onChanged(); } else { - return regionA_ == null ? - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionA_; + regionBuilder_.addMessage(builderForValue.build()); } + return this; } /** - * required .hbase.pb.RegionSpecifier region_a = 1; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> - getRegionAFieldBuilder() { - if (regionABuilder_ == null) { - regionABuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( - getRegionA(), - getParentForChildren(), - isClean()); - regionA_ = null; + public Builder addRegion( + int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { + if (regionBuilder_ == null) { + ensureRegionIsMutable(); + region_.add(index, builderForValue.build()); + onChanged(); + } else { + regionBuilder_.addMessage(index, builderForValue.build()); } - return regionABuilder_; - } - - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = null; - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; - /** - * required .hbase.pb.RegionSpecifier region_b = 2; - */ - public boolean hasRegionB() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return this; } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() { - if (regionBBuilder_ == null) { - return regionB_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; + public Builder addAllRegion( + java.lang.Iterable values) { + if (regionBuilder_ == null) { + ensureRegionIsMutable(); + org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, region_); + onChanged(); } else { - return regionBBuilder_.getMessage(); + regionBuilder_.addAllMessages(values); } + return this; } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder setRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - regionB_ = value; + public Builder clearRegion() { + if (regionBuilder_ == null) { + region_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { - regionBBuilder_.setMessage(value); + regionBuilder_.clear(); } - bitField0_ |= 0x00000002; return this; } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder setRegionB( - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionBBuilder_ == null) { - regionB_ = builderForValue.build(); + public Builder removeRegion(int index) { + if (regionBuilder_ == null) { + ensureRegionIsMutable(); + region_.remove(index); onChanged(); } else { - regionBBuilder_.setMessage(builderForValue.build()); + regionBuilder_.remove(index); } - bitField0_ |= 0x00000002; return this; } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder mergeRegionB(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && - regionB_ != null && - regionB_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { - regionB_ = - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionB_).mergeFrom(value).buildPartial(); - } else { - regionB_ = value; - } - onChanged(); - } else { - regionBBuilder_.mergeFrom(value); + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder( + int index) { + return getRegionFieldBuilder().getBuilder(index); + } + /** + * repeated .hbase.pb.RegionSpecifier region = 1; + */ + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( + int index) { + if (regionBuilder_ == null) { + return region_.get(index); } else { + return regionBuilder_.getMessageOrBuilder(index); } - bitField0_ |= 0x00000002; - return this; } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public Builder clearRegionB() { - if (regionBBuilder_ == null) { - regionB_ = null; - onChanged(); + public java.util.List + getRegionOrBuilderList() { + if (regionBuilder_ != null) { + return regionBuilder_.getMessageOrBuilderList(); } else { - regionBBuilder_.clear(); + return java.util.Collections.unmodifiableList(region_); } - bitField0_ = (bitField0_ & ~0x00000002); - return this; } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBBuilder() { - bitField0_ |= 0x00000002; - onChanged(); - return getRegionBFieldBuilder().getBuilder(); + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() { + return getRegionFieldBuilder().addBuilder( + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() { - if (regionBBuilder_ != null) { - return regionBBuilder_.getMessageOrBuilder(); - } else { - return regionB_ == null ? - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionB_; - } + public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder( + int index) { + return getRegionFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** - * required .hbase.pb.RegionSpecifier region_b = 2; + * repeated .hbase.pb.RegionSpecifier region = 1; */ - private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< + public java.util.List + getRegionBuilderList() { + return getRegionFieldBuilder().getBuilderList(); + } + private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> - getRegionBFieldBuilder() { - if (regionBBuilder_ == null) { - regionBBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3< + getRegionFieldBuilder() { + if (regionBuilder_ == null) { + regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( - getRegionB(), + region_, + ((bitField0_ & 0x00000001) == 0x00000001), getParentForChildren(), isClean()); - regionB_ = null; + region_ = null; } - return regionBBuilder_; + return regionBuilder_; } private boolean forcible_ ; @@ -6447,7 +6399,7 @@ public final class MasterProtos { * optional bool forcible = 3 [default = false]; */ public boolean hasForcible() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000002) == 0x00000002); } /** * optional bool forcible = 3 [default = false]; @@ -6459,7 +6411,7 @@ public final class MasterProtos { * optional bool forcible = 3 [default = false]; */ public Builder setForcible(boolean value) { - bitField0_ |= 0x00000004; + bitField0_ |= 0x00000002; forcible_ = value; onChanged(); return this; @@ -6468,7 +6420,7 @@ public final class MasterProtos { * optional bool forcible = 3 [default = false]; */ public Builder clearForcible() { - bitField0_ = (bitField0_ & ~0x00000004); + bitField0_ = (bitField0_ & ~0x00000002); forcible_ = false; onChanged(); return this; @@ -6479,7 +6431,7 @@ public final class MasterProtos { * optional uint64 nonce_group = 4 [default = 0]; */ public boolean hasNonceGroup() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000004) == 0x00000004); } /** * optional uint64 nonce_group = 4 [default = 0]; @@ -6491,7 +6443,7 @@ public final class MasterProtos { * optional uint64 nonce_group = 4 [default = 0]; */ public Builder setNonceGroup(long value) { - bitField0_ |= 0x00000008; + bitField0_ |= 0x00000004; nonceGroup_ = value; onChanged(); return this; @@ -6500,7 +6452,7 @@ public final class MasterProtos { * optional uint64 nonce_group = 4 [default = 0]; */ public Builder clearNonceGroup() { - bitField0_ = (bitField0_ & ~0x00000008); + bitField0_ = (bitField0_ & ~0x00000004); nonceGroup_ = 0L; onChanged(); return this; @@ -6511,7 +6463,7 @@ public final class MasterProtos { * optional uint64 nonce = 5 [default = 0]; */ public boolean hasNonce() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000008) == 0x00000008); } /** * optional uint64 nonce = 5 [default = 0]; @@ -6523,7 +6475,7 @@ public final class MasterProtos { * optional uint64 nonce = 5 [default = 0]; */ public Builder setNonce(long value) { - bitField0_ |= 0x00000010; + bitField0_ |= 0x00000008; nonce_ = value; onChanged(); return this; @@ -6532,7 +6484,7 @@ public final class MasterProtos { * optional uint64 nonce = 5 [default = 0]; */ public Builder clearNonce() { - bitField0_ = (bitField0_ & ~0x00000010); + bitField0_ = (bitField0_ & ~0x00000008); nonce_ = 0L; onChanged(); return this; @@ -6548,46 +6500,46 @@ public final class MasterProtos { } - // @@protoc_insertion_point(builder_scope:hbase.pb.DispatchMergingRegionsRequest) + // @@protoc_insertion_point(builder_scope:hbase.pb.MergeTableRegionsRequest) } - // @@protoc_insertion_point(class_scope:hbase.pb.DispatchMergingRegionsRequest) - private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest DEFAULT_INSTANCE; + // @@protoc_insertion_point(class_scope:hbase.pb.MergeTableRegionsRequest) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest DEFAULT_INSTANCE; static { - DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(); } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstance() { + public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } - @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser - PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser() { - public DispatchMergingRegionsRequest parsePartialFrom( + @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser + PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser() { + public MergeTableRegionsRequest parsePartialFrom( org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return new DispatchMergingRegionsRequest(input, extensionRegistry); + return new MergeTableRegionsRequest(input, extensionRegistry); } }; - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser parser() { + public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser getParserForType() { + public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser getParserForType() { return PARSER; } - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstanceForType() { + public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } - public interface DispatchMergingRegionsResponseOrBuilder extends - // @@protoc_insertion_point(interface_extends:hbase.pb.DispatchMergingRegionsResponse) + public interface MergeTableRegionsResponseOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsResponse) org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { /** @@ -6600,17 +6552,17 @@ public final class MasterProtos { long getProcId(); } /** - * Protobuf type {@code hbase.pb.DispatchMergingRegionsResponse} + * Protobuf type {@code hbase.pb.MergeTableRegionsResponse} */ - public static final class DispatchMergingRegionsResponse extends + public static final class MergeTableRegionsResponse extends org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:hbase.pb.DispatchMergingRegionsResponse) - DispatchMergingRegionsResponseOrBuilder { - // Use DispatchMergingRegionsResponse.newBuilder() to construct. - private DispatchMergingRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { + // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsResponse) + MergeTableRegionsResponseOrBuilder { + // Use MergeTableRegionsResponse.newBuilder() to construct. + private MergeTableRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } - private DispatchMergingRegionsResponse() { + private MergeTableRegionsResponse() { procId_ = 0L; } @@ -6619,7 +6571,7 @@ public final class MasterProtos { getUnknownFields() { return this.unknownFields; } - private DispatchMergingRegionsResponse( + private MergeTableRegionsResponse( org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { @@ -6661,14 +6613,14 @@ public final class MasterProtos { } public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor; + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_descriptor; } protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_fieldAccessorTable + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.Builder.class); + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.Builder.class); } private int bitField0_; @@ -6725,1508 +6677,10 @@ public final class MasterProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse)) { + if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse)) { return super.equals(obj); } - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) obj; - - boolean result = true; - result = result && (hasProcId() == other.hasProcId()); - if (hasProcId()) { - result = result && (getProcId() - == other.getProcId()); - } - result = result && unknownFields.equals(other.unknownFields); - return result; - } - - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptor().hashCode(); - if (hasProcId()) { - hash = (37 * hash) + PROC_ID_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong( - getProcId()); - } - hash = (29 * hash) + unknownFields.hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(byte[] data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( - byte[] data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseDelimitedFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder() { - return DEFAULT_INSTANCE.toBuilder(); - } - public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse prototype) { - return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { - return this == DEFAULT_INSTANCE - ? new Builder() : new Builder().mergeFrom(this); - } - - @java.lang.Override - protected Builder newBuilderForType( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.DispatchMergingRegionsResponse} - */ - public static final class Builder extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder implements - // @@protoc_insertion_point(builder_implements:hbase.pb.DispatchMergingRegionsResponse) - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponseOrBuilder { - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .alwaysUseFieldBuilders) { - } - } - public Builder clear() { - super.clear(); - procId_ = 0L; - bitField0_ = (bitField0_ & ~0x00000001); - return this; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstanceForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse build() { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse buildPartial() { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.procId_ = procId_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder clone() { - return (Builder) super.clone(); - } - public Builder setField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.setField(field, value); - } - public Builder clearField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) { - return (Builder) super.clearField(field); - } - public Builder clearOneof( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) { - return (Builder) super.clearOneof(oneof); - } - public Builder setRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - int index, Object value) { - return (Builder) super.setRepeatedField(field, index, value); - } - public Builder addRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.addRepeatedField(field, value); - } - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) { - return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other) { - if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance()) return this; - if (other.hasProcId()) { - setProcId(other.getProcId()); - } - this.mergeUnknownFields(other.unknownFields); - onChanged(); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) e.getUnfinishedMessage(); - throw e.unwrapIOException(); - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - private long procId_ ; - /** - * optional uint64 proc_id = 1; - */ - public boolean hasProcId() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * optional uint64 proc_id = 1; - */ - public long getProcId() { - return procId_; - } - /** - * optional uint64 proc_id = 1; - */ - public Builder setProcId(long value) { - bitField0_ |= 0x00000001; - procId_ = value; - onChanged(); - return this; - } - /** - * optional uint64 proc_id = 1; - */ - public Builder clearProcId() { - bitField0_ = (bitField0_ & ~0x00000001); - procId_ = 0L; - onChanged(); - return this; - } - public final Builder setUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.setUnknownFields(unknownFields); - } - - public final Builder mergeUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.mergeUnknownFields(unknownFields); - } - - - // @@protoc_insertion_point(builder_scope:hbase.pb.DispatchMergingRegionsResponse) - } - - // @@protoc_insertion_point(class_scope:hbase.pb.DispatchMergingRegionsResponse) - private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse DEFAULT_INSTANCE; - static { - DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse(); - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstance() { - return DEFAULT_INSTANCE; - } - - @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser - PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser() { - public DispatchMergingRegionsResponse parsePartialFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return new DispatchMergingRegionsResponse(input, extensionRegistry); - } - }; - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser parser() { - return PARSER; - } - - @java.lang.Override - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstanceForType() { - return DEFAULT_INSTANCE; - } - - } - - public interface MergeTableRegionsRequestOrBuilder extends - // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsRequest) - org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { - - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - java.util.List - getRegionList(); - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index); - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - int getRegionCount(); - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - java.util.List - getRegionOrBuilderList(); - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( - int index); - - /** - * optional bool forcible = 3 [default = false]; - */ - boolean hasForcible(); - /** - * optional bool forcible = 3 [default = false]; - */ - boolean getForcible(); - - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - boolean hasNonceGroup(); - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - long getNonceGroup(); - - /** - * optional uint64 nonce = 5 [default = 0]; - */ - boolean hasNonce(); - /** - * optional uint64 nonce = 5 [default = 0]; - */ - long getNonce(); - } - /** - *
-   **
-   * Merging the specified regions in a table.
-   * 
- * - * Protobuf type {@code hbase.pb.MergeTableRegionsRequest} - */ - public static final class MergeTableRegionsRequest extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsRequest) - MergeTableRegionsRequestOrBuilder { - // Use MergeTableRegionsRequest.newBuilder() to construct. - private MergeTableRegionsRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { - super(builder); - } - private MergeTableRegionsRequest() { - region_ = java.util.Collections.emptyList(); - forcible_ = false; - nonceGroup_ = 0L; - nonce_ = 0L; - } - - @java.lang.Override - public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private MergeTableRegionsRequest( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - this(); - int mutable_bitField0_ = 0; - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - region_ = new java.util.ArrayList(); - mutable_bitField0_ |= 0x00000001; - } - region_.add( - input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry)); - break; - } - case 24: { - bitField0_ |= 0x00000001; - forcible_ = input.readBool(); - break; - } - case 32: { - bitField0_ |= 0x00000002; - nonceGroup_ = input.readUInt64(); - break; - } - case 40: { - bitField0_ |= 0x00000004; - nonce_ = input.readUInt64(); - break; - } - } - } - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( - e).setUnfinishedMessage(this); - } finally { - if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - region_ = java.util.Collections.unmodifiableList(region_); - } - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class); - } - - private int bitField0_; - public static final int REGION_FIELD_NUMBER = 1; - private java.util.List region_; - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public java.util.List getRegionList() { - return region_; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public java.util.List - getRegionOrBuilderList() { - return region_; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public int getRegionCount() { - return region_.size(); - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { - return region_.get(index); - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( - int index) { - return region_.get(index); - } - - public static final int FORCIBLE_FIELD_NUMBER = 3; - private boolean forcible_; - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean hasForcible() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean getForcible() { - return forcible_; - } - - public static final int NONCE_GROUP_FIELD_NUMBER = 4; - private long nonceGroup_; - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - public boolean hasNonceGroup() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - public long getNonceGroup() { - return nonceGroup_; - } - - public static final int NONCE_FIELD_NUMBER = 5; - private long nonce_; - /** - * optional uint64 nonce = 5 [default = 0]; - */ - public boolean hasNonce() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional uint64 nonce = 5 [default = 0]; - */ - public long getNonce() { - return nonce_; - } - - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized == 1) return true; - if (isInitialized == 0) return false; - - for (int i = 0; i < getRegionCount(); i++) { - if (!getRegion(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - for (int i = 0; i < region_.size(); i++) { - output.writeMessage(1, region_.get(i)); - } - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBool(3, forcible_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeUInt64(4, nonceGroup_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeUInt64(5, nonce_); - } - unknownFields.writeTo(output); - } - - public int getSerializedSize() { - int size = memoizedSize; - if (size != -1) return size; - - size = 0; - for (int i = 0; i < region_.size(); i++) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeMessageSize(1, region_.get(i)); - } - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeBoolSize(3, forcible_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeUInt64Size(4, nonceGroup_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeUInt64Size(5, nonce_); - } - size += unknownFields.getSerializedSize(); - memoizedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) obj; - - boolean result = true; - result = result && getRegionList() - .equals(other.getRegionList()); - result = result && (hasForcible() == other.hasForcible()); - if (hasForcible()) { - result = result && (getForcible() - == other.getForcible()); - } - result = result && (hasNonceGroup() == other.hasNonceGroup()); - if (hasNonceGroup()) { - result = result && (getNonceGroup() - == other.getNonceGroup()); - } - result = result && (hasNonce() == other.hasNonce()); - if (hasNonce()) { - result = result && (getNonce() - == other.getNonce()); - } - result = result && unknownFields.equals(other.unknownFields); - return result; - } - - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptor().hashCode(); - if (getRegionCount() > 0) { - hash = (37 * hash) + REGION_FIELD_NUMBER; - hash = (53 * hash) + getRegionList().hashCode(); - } - if (hasForcible()) { - hash = (37 * hash) + FORCIBLE_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean( - getForcible()); - } - if (hasNonceGroup()) { - hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong( - getNonceGroup()); - } - if (hasNonce()) { - hash = (37 * hash) + NONCE_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong( - getNonce()); - } - hash = (29 * hash) + unknownFields.hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(byte[] data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( - byte[] data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder() { - return DEFAULT_INSTANCE.toBuilder(); - } - public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest prototype) { - return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { - return this == DEFAULT_INSTANCE - ? new Builder() : new Builder().mergeFrom(this); - } - - @java.lang.Override - protected Builder newBuilderForType( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - *
-     **
-     * Merging the specified regions in a table.
-     * 
- * - * Protobuf type {@code hbase.pb.MergeTableRegionsRequest} - */ - public static final class Builder extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder implements - // @@protoc_insertion_point(builder_implements:hbase.pb.MergeTableRegionsRequest) - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequestOrBuilder { - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .alwaysUseFieldBuilders) { - getRegionFieldBuilder(); - } - } - public Builder clear() { - super.clear(); - if (regionBuilder_ == null) { - region_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000001); - } else { - regionBuilder_.clear(); - } - forcible_ = false; - bitField0_ = (bitField0_ & ~0x00000002); - nonceGroup_ = 0L; - bitField0_ = (bitField0_ & ~0x00000004); - nonce_ = 0L; - bitField0_ = (bitField0_ & ~0x00000008); - return this; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest build() { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest buildPartial() { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (regionBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001)) { - region_ = java.util.Collections.unmodifiableList(region_); - bitField0_ = (bitField0_ & ~0x00000001); - } - result.region_ = region_; - } else { - result.region_ = regionBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000001; - } - result.forcible_ = forcible_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000002; - } - result.nonceGroup_ = nonceGroup_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000004; - } - result.nonce_ = nonce_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder clone() { - return (Builder) super.clone(); - } - public Builder setField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.setField(field, value); - } - public Builder clearField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) { - return (Builder) super.clearField(field); - } - public Builder clearOneof( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) { - return (Builder) super.clearOneof(oneof); - } - public Builder setRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - int index, Object value) { - return (Builder) super.setRepeatedField(field, index, value); - } - public Builder addRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.addRepeatedField(field, value); - } - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) { - return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other) { - if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance()) return this; - if (regionBuilder_ == null) { - if (!other.region_.isEmpty()) { - if (region_.isEmpty()) { - region_ = other.region_; - bitField0_ = (bitField0_ & ~0x00000001); - } else { - ensureRegionIsMutable(); - region_.addAll(other.region_); - } - onChanged(); - } - } else { - if (!other.region_.isEmpty()) { - if (regionBuilder_.isEmpty()) { - regionBuilder_.dispose(); - regionBuilder_ = null; - region_ = other.region_; - bitField0_ = (bitField0_ & ~0x00000001); - regionBuilder_ = - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? - getRegionFieldBuilder() : null; - } else { - regionBuilder_.addAllMessages(other.region_); - } - } - } - if (other.hasForcible()) { - setForcible(other.getForcible()); - } - if (other.hasNonceGroup()) { - setNonceGroup(other.getNonceGroup()); - } - if (other.hasNonce()) { - setNonce(other.getNonce()); - } - this.mergeUnknownFields(other.unknownFields); - onChanged(); - return this; - } - - public final boolean isInitialized() { - for (int i = 0; i < getRegionCount(); i++) { - if (!getRegion(i).isInitialized()) { - return false; - } - } - return true; - } - - public Builder mergeFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) e.getUnfinishedMessage(); - throw e.unwrapIOException(); - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - private java.util.List region_ = - java.util.Collections.emptyList(); - private void ensureRegionIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { - region_ = new java.util.ArrayList(region_); - bitField0_ |= 0x00000001; - } - } - - private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; - - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public java.util.List getRegionList() { - if (regionBuilder_ == null) { - return java.util.Collections.unmodifiableList(region_); - } else { - return regionBuilder_.getMessageList(); - } - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public int getRegionCount() { - if (regionBuilder_ == null) { - return region_.size(); - } else { - return regionBuilder_.getCount(); - } - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { - if (regionBuilder_ == null) { - return region_.get(index); - } else { - return regionBuilder_.getMessage(index); - } - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder setRegion( - int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureRegionIsMutable(); - region_.set(index, value); - onChanged(); - } else { - regionBuilder_.setMessage(index, value); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder setRegion( - int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionBuilder_ == null) { - ensureRegionIsMutable(); - region_.set(index, builderForValue.build()); - onChanged(); - } else { - regionBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureRegionIsMutable(); - region_.add(value); - onChanged(); - } else { - regionBuilder_.addMessage(value); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder addRegion( - int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureRegionIsMutable(); - region_.add(index, value); - onChanged(); - } else { - regionBuilder_.addMessage(index, value); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder addRegion( - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionBuilder_ == null) { - ensureRegionIsMutable(); - region_.add(builderForValue.build()); - onChanged(); - } else { - regionBuilder_.addMessage(builderForValue.build()); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder addRegion( - int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionBuilder_ == null) { - ensureRegionIsMutable(); - region_.add(index, builderForValue.build()); - onChanged(); - } else { - regionBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder addAllRegion( - java.lang.Iterable values) { - if (regionBuilder_ == null) { - ensureRegionIsMutable(); - org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll( - values, region_); - onChanged(); - } else { - regionBuilder_.addAllMessages(values); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder clearRegion() { - if (regionBuilder_ == null) { - region_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000001); - onChanged(); - } else { - regionBuilder_.clear(); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public Builder removeRegion(int index) { - if (regionBuilder_ == null) { - ensureRegionIsMutable(); - region_.remove(index); - onChanged(); - } else { - regionBuilder_.remove(index); - } - return this; - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder( - int index) { - return getRegionFieldBuilder().getBuilder(index); - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( - int index) { - if (regionBuilder_ == null) { - return region_.get(index); } else { - return regionBuilder_.getMessageOrBuilder(index); - } - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public java.util.List - getRegionOrBuilderList() { - if (regionBuilder_ != null) { - return regionBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(region_); - } - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() { - return getRegionFieldBuilder().addBuilder( - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder( - int index) { - return getRegionFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); - } - /** - * repeated .hbase.pb.RegionSpecifier region = 1; - */ - public java.util.List - getRegionBuilderList() { - return getRegionFieldBuilder().getBuilderList(); - } - private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> - getRegionFieldBuilder() { - if (regionBuilder_ == null) { - regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( - region_, - ((bitField0_ & 0x00000001) == 0x00000001), - getParentForChildren(), - isClean()); - region_ = null; - } - return regionBuilder_; - } - - private boolean forcible_ ; - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean hasForcible() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * optional bool forcible = 3 [default = false]; - */ - public boolean getForcible() { - return forcible_; - } - /** - * optional bool forcible = 3 [default = false]; - */ - public Builder setForcible(boolean value) { - bitField0_ |= 0x00000002; - forcible_ = value; - onChanged(); - return this; - } - /** - * optional bool forcible = 3 [default = false]; - */ - public Builder clearForcible() { - bitField0_ = (bitField0_ & ~0x00000002); - forcible_ = false; - onChanged(); - return this; - } - - private long nonceGroup_ ; - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - public boolean hasNonceGroup() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - public long getNonceGroup() { - return nonceGroup_; - } - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - public Builder setNonceGroup(long value) { - bitField0_ |= 0x00000004; - nonceGroup_ = value; - onChanged(); - return this; - } - /** - * optional uint64 nonce_group = 4 [default = 0]; - */ - public Builder clearNonceGroup() { - bitField0_ = (bitField0_ & ~0x00000004); - nonceGroup_ = 0L; - onChanged(); - return this; - } - - private long nonce_ ; - /** - * optional uint64 nonce = 5 [default = 0]; - */ - public boolean hasNonce() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - * optional uint64 nonce = 5 [default = 0]; - */ - public long getNonce() { - return nonce_; - } - /** - * optional uint64 nonce = 5 [default = 0]; - */ - public Builder setNonce(long value) { - bitField0_ |= 0x00000008; - nonce_ = value; - onChanged(); - return this; - } - /** - * optional uint64 nonce = 5 [default = 0]; - */ - public Builder clearNonce() { - bitField0_ = (bitField0_ & ~0x00000008); - nonce_ = 0L; - onChanged(); - return this; - } - public final Builder setUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.setUnknownFields(unknownFields); - } - - public final Builder mergeUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return super.mergeUnknownFields(unknownFields); - } - - - // @@protoc_insertion_point(builder_scope:hbase.pb.MergeTableRegionsRequest) - } - - // @@protoc_insertion_point(class_scope:hbase.pb.MergeTableRegionsRequest) - private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest DEFAULT_INSTANCE; - static { - DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(); - } - - public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstance() { - return DEFAULT_INSTANCE; - } - - @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser - PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser() { - public MergeTableRegionsRequest parsePartialFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return new MergeTableRegionsRequest(input, extensionRegistry); - } - }; - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser parser() { - return PARSER; - } - - @java.lang.Override - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() { - return DEFAULT_INSTANCE; - } - - } - - public interface MergeTableRegionsResponseOrBuilder extends - // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsResponse) - org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { - - /** - * optional uint64 proc_id = 1; - */ - boolean hasProcId(); - /** - * optional uint64 proc_id = 1; - */ - long getProcId(); - } - /** - * Protobuf type {@code hbase.pb.MergeTableRegionsResponse} - */ - public static final class MergeTableRegionsResponse extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsResponse) - MergeTableRegionsResponseOrBuilder { - // Use MergeTableRegionsResponse.newBuilder() to construct. - private MergeTableRegionsResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder builder) { - super(builder); - } - private MergeTableRegionsResponse() { - procId_ = 0L; - } - - @java.lang.Override - public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private MergeTableRegionsResponse( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - this(); - int mutable_bitField0_ = 0; - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields = - org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 8: { - bitField0_ |= 0x00000001; - procId_ = input.readUInt64(); - break; - } - } - } - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( - e).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.Builder.class); - } - - private int bitField0_; - public static final int PROC_ID_FIELD_NUMBER = 1; - private long procId_; - /** - * optional uint64 proc_id = 1; - */ - public boolean hasProcId() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * optional uint64 proc_id = 1; - */ - public long getProcId() { - return procId_; - } - - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized == 1) return true; - if (isInitialized == 0) return false; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeUInt64(1, procId_); - } - unknownFields.writeTo(output); - } - - public int getSerializedSize() { - int size = memoizedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeUInt64Size(1, procId_); - } - size += unknownFields.getSerializedSize(); - memoizedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) obj; + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) obj; boolean result = true; result = result && (hasProcId() == other.hasProcId()); @@ -74504,18 +72958,6 @@ public final class MasterProtos { /** *
-       ** Master dispatch merging the regions 
-       * 
- * - * rpc DispatchMergingRegions(.hbase.pb.DispatchMergingRegionsRequest) returns (.hbase.pb.DispatchMergingRegionsResponse); - */ - public abstract void dispatchMergingRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done); - - /** - *
        ** Move the region region to the destination server. 
        * 
* @@ -75404,14 +73846,6 @@ public final class MasterProtos { } @java.lang.Override - public void dispatchMergingRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { - impl.dispatchMergingRegions(controller, request, done); - } - - @java.lang.Override public void moveRegion( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request, @@ -75978,138 +74412,136 @@ public final class MasterProtos { case 7: return impl.modifyColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest)request); case 8: - return impl.dispatchMergingRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)request); - case 9: return impl.moveRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest)request); - case 10: + case 9: return impl.mergeTableRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)request); - case 11: + case 10: return impl.assignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest)request); - case 12: + case 11: return impl.unassignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest)request); - case 13: + case 12: return impl.offlineRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest)request); - case 14: + case 13: return impl.splitRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest)request); - case 15: + case 14: return impl.deleteTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest)request); - case 16: + case 15: return impl.truncateTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest)request); - case 17: + case 16: return impl.enableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest)request); - case 18: + case 17: return impl.disableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest)request); - case 19: + case 18: return impl.modifyTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest)request); - case 20: + case 19: return impl.createTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest)request); - case 21: + case 20: return impl.shutdown(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest)request); - case 22: + case 21: return impl.stopMaster(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest)request); - case 23: + case 22: return impl.isMasterInMaintenanceMode(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest)request); - case 24: + case 23: return impl.balance(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest)request); - case 25: + case 24: return impl.setBalancerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest)request); - case 26: + case 25: return impl.isBalancerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest)request); - case 27: + case 26: return impl.setSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)request); - case 28: + case 27: return impl.isSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)request); - case 29: + case 28: return impl.normalize(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest)request); - case 30: + case 29: return impl.setNormalizerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)request); - case 31: + case 30: return impl.isNormalizerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest)request); - case 32: + case 31: return impl.runCatalogScan(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest)request); - case 33: + case 32: return impl.enableCatalogJanitor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest)request); - case 34: + case 33: return impl.isCatalogJanitorEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest)request); - case 35: + case 34: return impl.runCleanerChore(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest)request); - case 36: + case 35: return impl.setCleanerChoreRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest)request); - case 37: + case 36: return impl.isCleanerChoreEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest)request); - case 38: + case 37: return impl.execMasterService(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest)request); - case 39: + case 38: return impl.snapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest)request); - case 40: + case 39: return impl.getCompletedSnapshots(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest)request); - case 41: + case 40: return impl.deleteSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest)request); - case 42: + case 41: return impl.isSnapshotDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest)request); - case 43: + case 42: return impl.restoreSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest)request); - case 44: + case 43: return impl.execProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request); - case 45: + case 44: return impl.execProcedureWithRet(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request); - case 46: + case 45: return impl.isProcedureDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest)request); - case 47: + case 46: return impl.modifyNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest)request); - case 48: + case 47: return impl.createNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest)request); - case 49: + case 48: return impl.deleteNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest)request); - case 50: + case 49: return impl.getNamespaceDescriptor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest)request); - case 51: + case 50: return impl.listNamespaceDescriptors(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest)request); - case 52: + case 51: return impl.listTableDescriptorsByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest)request); - case 53: + case 52: return impl.listTableNamesByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest)request); - case 54: + case 53: return impl.getTableState(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest)request); - case 55: + case 54: return impl.setQuota(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest)request); - case 56: + case 55: return impl.getLastMajorCompactionTimestamp(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest)request); - case 57: + case 56: return impl.getLastMajorCompactionTimestampForRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest)request); - case 58: + case 57: return impl.getProcedureResult(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest)request); - case 59: + case 58: return impl.getSecurityCapabilities(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest)request); - case 60: + case 59: return impl.abortProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest)request); - case 61: + case 60: return impl.listProcedures(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresRequest)request); - case 62: + case 61: return impl.listLocks(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksRequest)request); - case 63: + case 62: return impl.addReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest)request); - case 64: + case 63: return impl.removeReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest)request); - case 65: + case 64: return impl.enableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest)request); - case 66: + case 65: return impl.disableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)request); - case 67: + case 66: return impl.getReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)request); - case 68: + case 67: return impl.updateReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)request); - case 69: + case 68: return impl.listReplicationPeers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest)request); - case 70: + case 69: return impl.listDrainingRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest)request); - case 71: + case 70: return impl.drainRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest)request); - case 72: + case 71: return impl.removeDrainFromRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest)request); - case 73: + case 72: return impl.getSpaceQuotaRegionSizes(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest)request); - case 74: + case 73: return impl.getQuotaStates(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest)request); default: throw new java.lang.AssertionError("Can't get here."); @@ -76142,138 +74574,136 @@ public final class MasterProtos { case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance(); case 8: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance(); - case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance(); - case 10: + case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); - case 11: + case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance(); - case 12: + case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance(); - case 13: + case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance(); - case 14: + case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.getDefaultInstance(); - case 15: + case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.getDefaultInstance(); - case 16: + case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.getDefaultInstance(); - case 17: + case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.getDefaultInstance(); - case 18: + case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.getDefaultInstance(); - case 19: + case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.getDefaultInstance(); - case 20: + case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.getDefaultInstance(); - case 21: + case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.getDefaultInstance(); - case 22: + case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.getDefaultInstance(); - case 23: + case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.getDefaultInstance(); - case 24: + case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.getDefaultInstance(); - case 25: + case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.getDefaultInstance(); - case 26: + case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.getDefaultInstance(); - case 27: + case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance(); - case 28: + case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance(); - case 29: + case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance(); - case 30: + case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance(); - case 31: + case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.getDefaultInstance(); - case 32: + case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.getDefaultInstance(); - case 33: + case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.getDefaultInstance(); - case 34: + case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.getDefaultInstance(); - case 35: + case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.getDefaultInstance(); - case 36: + case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.getDefaultInstance(); - case 37: + case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.getDefaultInstance(); - case 38: + case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest.getDefaultInstance(); - case 39: + case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.getDefaultInstance(); - case 40: + case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.getDefaultInstance(); - case 41: + case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.getDefaultInstance(); - case 42: + case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.getDefaultInstance(); - case 43: + case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.getDefaultInstance(); + case 43: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 45: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); - case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.getDefaultInstance(); - case 47: + case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.getDefaultInstance(); - case 48: + case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.getDefaultInstance(); - case 49: + case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.getDefaultInstance(); - case 50: + case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.getDefaultInstance(); - case 51: + case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.getDefaultInstance(); - case 52: + case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.getDefaultInstance(); - case 53: + case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.getDefaultInstance(); - case 54: + case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.getDefaultInstance(); - case 55: + case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.getDefaultInstance(); - case 56: + case 55: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.getDefaultInstance(); - case 57: + case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.getDefaultInstance(); - case 58: + case 57: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.getDefaultInstance(); - case 59: + case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.getDefaultInstance(); - case 60: + case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.getDefaultInstance(); - case 61: + case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresRequest.getDefaultInstance(); - case 62: + case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksRequest.getDefaultInstance(); - case 63: + case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.getDefaultInstance(); - case 64: + case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.getDefaultInstance(); - case 65: + case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.getDefaultInstance(); - case 66: + case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); - case 67: + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); - case 68: + case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); - case 69: + case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.getDefaultInstance(); - case 70: + case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest.getDefaultInstance(); - case 71: + case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest.getDefaultInstance(); - case 72: + case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest.getDefaultInstance(); - case 73: + case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest.getDefaultInstance(); - case 74: + case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -76306,138 +74736,136 @@ public final class MasterProtos { case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance(); case 8: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance(); - case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(); - case 10: + case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(); - case 11: + case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(); - case 12: + case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(); - case 13: + case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(); - case 14: + case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(); - case 15: + case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(); - case 16: + case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(); - case 17: + case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(); - case 18: + case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(); - case 19: + case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(); - case 20: + case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(); - case 21: + case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(); - case 22: + case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(); - case 23: + case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(); - case 24: + case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(); - case 25: + case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(); - case 26: + case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(); - case 27: + case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(); - case 28: + case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(); - case 29: + case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(); - case 30: + case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(); - case 31: + case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(); - case 32: + case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(); - case 33: + case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(); - case 34: + case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(); - case 35: + case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(); - case 36: + case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(); - case 37: + case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(); - case 38: + case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance(); - case 39: + case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(); - case 40: + case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(); - case 41: + case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(); - case 42: + case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(); - case 43: + case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(); + case 43: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 45: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); - case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(); - case 47: + case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(); - case 48: + case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(); - case 49: + case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(); - case 50: + case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(); - case 51: + case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(); - case 52: + case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(); - case 53: + case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(); - case 54: + case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); - case 55: + case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(); + case 55: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 57: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); - case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(); - case 59: + case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(); - case 60: + case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(); - case 61: + case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresResponse.getDefaultInstance(); - case 62: + case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksResponse.getDefaultInstance(); - case 63: + case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(); - case 64: + case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(); - case 65: + case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(); - case 66: + case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); - case 67: + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); - case 68: + case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); - case 69: + case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(); - case 70: + case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(); - case 71: + case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(); - case 72: + case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance(); - case 73: + case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance(); - case 74: + case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -76545,18 +74973,6 @@ public final class MasterProtos { /** *
-     ** Master dispatch merging the regions 
-     * 
- * - * rpc DispatchMergingRegions(.hbase.pb.DispatchMergingRegionsRequest) returns (.hbase.pb.DispatchMergingRegionsResponse); - */ - public abstract void dispatchMergingRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done); - - /** - *
      ** Move the region region to the destination server. 
      * 
* @@ -77438,336 +75854,331 @@ public final class MasterProtos { done)); return; case 8: - this.dispatchMergingRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( - done)); - return; - case 9: this.moveRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 10: + case 9: this.mergeTableRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 11: + case 10: this.assignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 12: + case 11: this.unassignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 13: + case 12: this.offlineRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 14: + case 13: this.splitRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 15: + case 14: this.deleteTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 16: + case 15: this.truncateTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 17: + case 16: this.enableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 18: + case 17: this.disableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 19: + case 18: this.modifyTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 20: + case 19: this.createTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 21: + case 20: this.shutdown(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 22: + case 21: this.stopMaster(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 23: + case 22: this.isMasterInMaintenanceMode(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 24: + case 23: this.balance(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 25: + case 24: this.setBalancerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 26: + case 25: this.isBalancerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 27: + case 26: this.setSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 28: + case 27: this.isSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 29: + case 28: this.normalize(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 30: + case 29: this.setNormalizerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 31: + case 30: this.isNormalizerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 32: + case 31: this.runCatalogScan(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 33: + case 32: this.enableCatalogJanitor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 34: + case 33: this.isCatalogJanitorEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 35: + case 34: this.runCleanerChore(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 36: + case 35: this.setCleanerChoreRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 37: + case 36: this.isCleanerChoreEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 38: + case 37: this.execMasterService(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 39: + case 38: this.snapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 40: + case 39: this.getCompletedSnapshots(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 41: + case 40: this.deleteSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 42: + case 41: this.isSnapshotDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 43: + case 42: this.restoreSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 44: + case 43: this.execProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 45: + case 44: this.execProcedureWithRet(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 46: + case 45: this.isProcedureDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 47: + case 46: this.modifyNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 48: + case 47: this.createNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 49: + case 48: this.deleteNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 50: + case 49: this.getNamespaceDescriptor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 51: + case 50: this.listNamespaceDescriptors(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 52: + case 51: this.listTableDescriptorsByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 53: + case 52: this.listTableNamesByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 54: + case 53: this.getTableState(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 55: + case 54: this.setQuota(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 56: + case 55: this.getLastMajorCompactionTimestamp(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 57: + case 56: this.getLastMajorCompactionTimestampForRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 58: + case 57: this.getProcedureResult(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 59: + case 58: this.getSecurityCapabilities(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 60: + case 59: this.abortProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 61: + case 60: this.listProcedures(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 62: + case 61: this.listLocks(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 63: + case 62: this.addReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 64: + case 63: this.removeReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 65: + case 64: this.enableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 66: + case 65: this.disableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 67: + case 66: this.getReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 68: + case 67: this.updateReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 69: + case 68: this.listReplicationPeers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 70: + case 69: this.listDrainingRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 71: + case 70: this.drainRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 72: + case 71: this.removeDrainFromRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 73: + case 72: this.getSpaceQuotaRegionSizes(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); return; - case 74: + case 73: this.getQuotaStates(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.specializeCallback( done)); @@ -77803,138 +76214,136 @@ public final class MasterProtos { case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance(); case 8: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance(); - case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance(); - case 10: + case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); - case 11: + case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance(); - case 12: + case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance(); - case 13: + case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance(); - case 14: + case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.getDefaultInstance(); - case 15: + case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.getDefaultInstance(); - case 16: + case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.getDefaultInstance(); - case 17: + case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.getDefaultInstance(); - case 18: + case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.getDefaultInstance(); - case 19: + case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.getDefaultInstance(); - case 20: + case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.getDefaultInstance(); - case 21: + case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.getDefaultInstance(); - case 22: + case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.getDefaultInstance(); - case 23: + case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.getDefaultInstance(); - case 24: + case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.getDefaultInstance(); - case 25: + case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.getDefaultInstance(); - case 26: + case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.getDefaultInstance(); - case 27: + case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance(); - case 28: + case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance(); - case 29: + case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance(); - case 30: + case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance(); - case 31: + case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.getDefaultInstance(); - case 32: + case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.getDefaultInstance(); - case 33: + case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.getDefaultInstance(); - case 34: + case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.getDefaultInstance(); - case 35: + case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.getDefaultInstance(); - case 36: + case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.getDefaultInstance(); - case 37: + case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.getDefaultInstance(); - case 38: + case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest.getDefaultInstance(); - case 39: + case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.getDefaultInstance(); - case 40: + case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.getDefaultInstance(); - case 41: + case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.getDefaultInstance(); - case 42: + case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.getDefaultInstance(); - case 43: + case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.getDefaultInstance(); + case 43: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 45: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); - case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.getDefaultInstance(); - case 47: + case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.getDefaultInstance(); - case 48: + case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.getDefaultInstance(); - case 49: + case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.getDefaultInstance(); - case 50: + case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.getDefaultInstance(); - case 51: + case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.getDefaultInstance(); - case 52: + case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.getDefaultInstance(); - case 53: + case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.getDefaultInstance(); - case 54: + case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.getDefaultInstance(); - case 55: + case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.getDefaultInstance(); - case 56: + case 55: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.getDefaultInstance(); - case 57: + case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.getDefaultInstance(); - case 58: + case 57: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.getDefaultInstance(); - case 59: + case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.getDefaultInstance(); - case 60: + case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.getDefaultInstance(); - case 61: + case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresRequest.getDefaultInstance(); - case 62: + case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksRequest.getDefaultInstance(); - case 63: + case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.getDefaultInstance(); - case 64: + case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.getDefaultInstance(); - case 65: + case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.getDefaultInstance(); - case 66: + case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); - case 67: + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); - case 68: + case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); - case 69: + case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.getDefaultInstance(); - case 70: + case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest.getDefaultInstance(); - case 71: + case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest.getDefaultInstance(); - case 72: + case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest.getDefaultInstance(); - case 73: + case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest.getDefaultInstance(); - case 74: + case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -77967,138 +76376,136 @@ public final class MasterProtos { case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance(); case 8: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance(); - case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(); - case 10: + case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(); - case 11: + case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(); - case 12: + case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(); - case 13: + case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(); - case 14: + case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(); - case 15: + case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(); - case 16: + case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(); - case 17: + case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(); - case 18: + case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(); - case 19: + case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(); - case 20: + case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(); - case 21: + case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(); - case 22: + case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(); - case 23: + case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(); - case 24: + case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(); - case 25: + case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(); - case 26: + case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(); - case 27: + case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(); - case 28: + case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(); - case 29: + case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(); - case 30: + case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(); - case 31: + case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(); - case 32: + case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(); - case 33: + case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(); - case 34: + case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(); - case 35: + case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(); - case 36: + case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(); - case 37: + case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(); - case 38: + case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance(); - case 39: + case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(); - case 40: + case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(); - case 41: + case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(); - case 42: + case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(); - case 43: + case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(); + case 43: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 45: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); - case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(); - case 47: + case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(); - case 48: + case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(); - case 49: + case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(); - case 50: + case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(); - case 51: + case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(); - case 52: + case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(); - case 53: + case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(); - case 54: + case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); - case 55: + case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(); + case 55: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 57: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); - case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(); - case 59: + case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(); - case 60: + case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(); - case 61: + case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresResponse.getDefaultInstance(); - case 62: + case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksResponse.getDefaultInstance(); - case 63: + case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(); - case 64: + case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(); - case 65: + case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(); - case 66: + case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); - case 67: + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); - case 68: + case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); - case 69: + case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(); - case 70: + case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(); - case 71: + case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(); - case 72: + case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance(); - case 73: + case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance(); - case 74: + case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -78241,27 +76648,12 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance())); } - public void dispatchMergingRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest request, - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { - channel.callMethod( - getDescriptor().getMethods().get(8), - controller, - request, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance(), - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( - done, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance())); - } - public void moveRegion( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(9), + getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(), @@ -78276,7 +76668,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(10), + getDescriptor().getMethods().get(9), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(), @@ -78291,7 +76683,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(11), + getDescriptor().getMethods().get(10), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(), @@ -78306,7 +76698,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(12), + getDescriptor().getMethods().get(11), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(), @@ -78321,7 +76713,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(13), + getDescriptor().getMethods().get(12), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(), @@ -78336,7 +76728,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(14), + getDescriptor().getMethods().get(13), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(), @@ -78351,7 +76743,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(15), + getDescriptor().getMethods().get(14), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(), @@ -78366,7 +76758,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(16), + getDescriptor().getMethods().get(15), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(), @@ -78381,7 +76773,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(17), + getDescriptor().getMethods().get(16), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(), @@ -78396,7 +76788,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(18), + getDescriptor().getMethods().get(17), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(), @@ -78411,7 +76803,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(19), + getDescriptor().getMethods().get(18), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(), @@ -78426,7 +76818,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(20), + getDescriptor().getMethods().get(19), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(), @@ -78441,7 +76833,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(21), + getDescriptor().getMethods().get(20), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(), @@ -78456,7 +76848,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(22), + getDescriptor().getMethods().get(21), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(), @@ -78471,7 +76863,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(23), + getDescriptor().getMethods().get(22), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(), @@ -78486,7 +76878,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(24), + getDescriptor().getMethods().get(23), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(), @@ -78501,7 +76893,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(25), + getDescriptor().getMethods().get(24), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(), @@ -78516,7 +76908,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(26), + getDescriptor().getMethods().get(25), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(), @@ -78531,7 +76923,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(27), + getDescriptor().getMethods().get(26), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(), @@ -78546,7 +76938,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(28), + getDescriptor().getMethods().get(27), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(), @@ -78561,7 +76953,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(29), + getDescriptor().getMethods().get(28), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(), @@ -78576,7 +76968,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(30), + getDescriptor().getMethods().get(29), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(), @@ -78591,7 +76983,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(31), + getDescriptor().getMethods().get(30), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(), @@ -78606,7 +76998,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(32), + getDescriptor().getMethods().get(31), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(), @@ -78621,7 +77013,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(33), + getDescriptor().getMethods().get(32), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(), @@ -78636,7 +77028,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(34), + getDescriptor().getMethods().get(33), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(), @@ -78651,7 +77043,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(35), + getDescriptor().getMethods().get(34), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(), @@ -78666,7 +77058,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(36), + getDescriptor().getMethods().get(35), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(), @@ -78681,7 +77073,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(37), + getDescriptor().getMethods().get(36), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(), @@ -78696,7 +77088,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(38), + getDescriptor().getMethods().get(37), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance(), @@ -78711,7 +77103,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(39), + getDescriptor().getMethods().get(38), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(), @@ -78726,7 +77118,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(40), + getDescriptor().getMethods().get(39), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(), @@ -78741,7 +77133,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(41), + getDescriptor().getMethods().get(40), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(), @@ -78756,7 +77148,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(42), + getDescriptor().getMethods().get(41), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(), @@ -78771,7 +77163,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(43), + getDescriptor().getMethods().get(42), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(), @@ -78786,7 +77178,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(44), + getDescriptor().getMethods().get(43), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(), @@ -78801,7 +77193,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(45), + getDescriptor().getMethods().get(44), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(), @@ -78816,7 +77208,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(46), + getDescriptor().getMethods().get(45), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(), @@ -78831,7 +77223,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(47), + getDescriptor().getMethods().get(46), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(), @@ -78846,7 +77238,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(48), + getDescriptor().getMethods().get(47), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(), @@ -78861,7 +77253,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(49), + getDescriptor().getMethods().get(48), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(), @@ -78876,7 +77268,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(50), + getDescriptor().getMethods().get(49), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(), @@ -78891,7 +77283,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(51), + getDescriptor().getMethods().get(50), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(), @@ -78906,7 +77298,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(52), + getDescriptor().getMethods().get(51), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(), @@ -78921,7 +77313,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(53), + getDescriptor().getMethods().get(52), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(), @@ -78936,7 +77328,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(54), + getDescriptor().getMethods().get(53), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(), @@ -78951,7 +77343,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(55), + getDescriptor().getMethods().get(54), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(), @@ -78966,7 +77358,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(56), + getDescriptor().getMethods().get(55), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(), @@ -78981,7 +77373,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(57), + getDescriptor().getMethods().get(56), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(), @@ -78996,7 +77388,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(58), + getDescriptor().getMethods().get(57), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(), @@ -79011,7 +77403,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(59), + getDescriptor().getMethods().get(58), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(), @@ -79026,7 +77418,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(60), + getDescriptor().getMethods().get(59), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(), @@ -79041,7 +77433,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(61), + getDescriptor().getMethods().get(60), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresResponse.getDefaultInstance(), @@ -79056,7 +77448,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(62), + getDescriptor().getMethods().get(61), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksResponse.getDefaultInstance(), @@ -79071,7 +77463,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(63), + getDescriptor().getMethods().get(62), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(), @@ -79086,7 +77478,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(64), + getDescriptor().getMethods().get(63), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(), @@ -79101,7 +77493,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(65), + getDescriptor().getMethods().get(64), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(), @@ -79116,7 +77508,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(66), + getDescriptor().getMethods().get(65), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(), @@ -79131,7 +77523,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(67), + getDescriptor().getMethods().get(66), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(), @@ -79146,7 +77538,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(68), + getDescriptor().getMethods().get(67), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(), @@ -79161,7 +77553,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(69), + getDescriptor().getMethods().get(68), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(), @@ -79176,7 +77568,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(70), + getDescriptor().getMethods().get(69), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(), @@ -79191,7 +77583,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(71), + getDescriptor().getMethods().get(70), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(), @@ -79206,7 +77598,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(72), + getDescriptor().getMethods().get(71), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance(), @@ -79221,7 +77613,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(73), + getDescriptor().getMethods().get(72), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance(), @@ -79236,7 +77628,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback done) { channel.callMethod( - getDescriptor().getMethods().get(74), + getDescriptor().getMethods().get(73), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance(), @@ -79293,11 +77685,6 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse dispatchMergingRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest request) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse moveRegion( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request) @@ -79732,24 +78119,12 @@ public final class MasterProtos { } - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse dispatchMergingRegions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest request) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { - return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(8), - controller, - request, - org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance()); - } - - public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse moveRegion( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(9), + getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance()); @@ -79761,7 +78136,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(10), + getDescriptor().getMethods().get(9), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance()); @@ -79773,7 +78148,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(11), + getDescriptor().getMethods().get(10), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance()); @@ -79785,7 +78160,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(12), + getDescriptor().getMethods().get(11), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance()); @@ -79797,7 +78172,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(13), + getDescriptor().getMethods().get(12), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance()); @@ -79809,7 +78184,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(14), + getDescriptor().getMethods().get(13), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance()); @@ -79821,7 +78196,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(15), + getDescriptor().getMethods().get(14), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance()); @@ -79833,7 +78208,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(16), + getDescriptor().getMethods().get(15), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance()); @@ -79845,7 +78220,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(17), + getDescriptor().getMethods().get(16), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance()); @@ -79857,7 +78232,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(18), + getDescriptor().getMethods().get(17), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance()); @@ -79869,7 +78244,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(19), + getDescriptor().getMethods().get(18), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance()); @@ -79881,7 +78256,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(20), + getDescriptor().getMethods().get(19), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance()); @@ -79893,7 +78268,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(21), + getDescriptor().getMethods().get(20), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance()); @@ -79905,7 +78280,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(22), + getDescriptor().getMethods().get(21), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance()); @@ -79917,7 +78292,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(23), + getDescriptor().getMethods().get(22), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance()); @@ -79929,7 +78304,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(24), + getDescriptor().getMethods().get(23), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance()); @@ -79941,7 +78316,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(25), + getDescriptor().getMethods().get(24), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance()); @@ -79953,7 +78328,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(26), + getDescriptor().getMethods().get(25), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance()); @@ -79965,7 +78340,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(27), + getDescriptor().getMethods().get(26), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance()); @@ -79977,7 +78352,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(28), + getDescriptor().getMethods().get(27), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance()); @@ -79989,7 +78364,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(29), + getDescriptor().getMethods().get(28), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance()); @@ -80001,7 +78376,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(30), + getDescriptor().getMethods().get(29), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance()); @@ -80013,7 +78388,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(31), + getDescriptor().getMethods().get(30), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance()); @@ -80025,7 +78400,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(32), + getDescriptor().getMethods().get(31), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance()); @@ -80037,7 +78412,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(33), + getDescriptor().getMethods().get(32), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance()); @@ -80049,7 +78424,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(34), + getDescriptor().getMethods().get(33), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance()); @@ -80061,7 +78436,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(35), + getDescriptor().getMethods().get(34), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance()); @@ -80073,7 +78448,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(36), + getDescriptor().getMethods().get(35), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance()); @@ -80085,7 +78460,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(37), + getDescriptor().getMethods().get(36), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance()); @@ -80097,7 +78472,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(38), + getDescriptor().getMethods().get(37), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance()); @@ -80109,7 +78484,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(39), + getDescriptor().getMethods().get(38), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance()); @@ -80121,7 +78496,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(40), + getDescriptor().getMethods().get(39), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance()); @@ -80133,7 +78508,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(41), + getDescriptor().getMethods().get(40), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance()); @@ -80145,7 +78520,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(42), + getDescriptor().getMethods().get(41), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance()); @@ -80157,7 +78532,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(43), + getDescriptor().getMethods().get(42), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance()); @@ -80169,7 +78544,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(44), + getDescriptor().getMethods().get(43), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance()); @@ -80181,7 +78556,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(45), + getDescriptor().getMethods().get(44), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance()); @@ -80193,7 +78568,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(46), + getDescriptor().getMethods().get(45), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance()); @@ -80205,7 +78580,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(47), + getDescriptor().getMethods().get(46), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance()); @@ -80217,7 +78592,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(48), + getDescriptor().getMethods().get(47), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance()); @@ -80229,7 +78604,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(49), + getDescriptor().getMethods().get(48), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance()); @@ -80241,7 +78616,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(50), + getDescriptor().getMethods().get(49), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance()); @@ -80253,7 +78628,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(51), + getDescriptor().getMethods().get(50), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance()); @@ -80265,7 +78640,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(52), + getDescriptor().getMethods().get(51), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance()); @@ -80277,7 +78652,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(53), + getDescriptor().getMethods().get(52), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance()); @@ -80289,7 +78664,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(54), + getDescriptor().getMethods().get(53), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance()); @@ -80301,7 +78676,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(55), + getDescriptor().getMethods().get(54), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance()); @@ -80313,7 +78688,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(56), + getDescriptor().getMethods().get(55), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance()); @@ -80325,7 +78700,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(57), + getDescriptor().getMethods().get(56), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance()); @@ -80337,7 +78712,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(58), + getDescriptor().getMethods().get(57), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance()); @@ -80349,7 +78724,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(59), + getDescriptor().getMethods().get(58), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance()); @@ -80361,7 +78736,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(60), + getDescriptor().getMethods().get(59), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance()); @@ -80373,7 +78748,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(61), + getDescriptor().getMethods().get(60), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListProceduresResponse.getDefaultInstance()); @@ -80385,7 +78760,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(62), + getDescriptor().getMethods().get(61), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListLocksResponse.getDefaultInstance()); @@ -80397,7 +78772,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(63), + getDescriptor().getMethods().get(62), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance()); @@ -80409,7 +78784,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(64), + getDescriptor().getMethods().get(63), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance()); @@ -80421,7 +78796,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(65), + getDescriptor().getMethods().get(64), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance()); @@ -80433,7 +78808,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(66), + getDescriptor().getMethods().get(65), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance()); @@ -80445,7 +78820,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(67), + getDescriptor().getMethods().get(66), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance()); @@ -80457,7 +78832,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(68), + getDescriptor().getMethods().get(67), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance()); @@ -80469,7 +78844,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(69), + getDescriptor().getMethods().get(68), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance()); @@ -80481,7 +78856,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(70), + getDescriptor().getMethods().get(69), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance()); @@ -80493,7 +78868,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(71), + getDescriptor().getMethods().get(70), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance()); @@ -80505,7 +78880,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(72), + getDescriptor().getMethods().get(71), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance()); @@ -80517,7 +78892,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(73), + getDescriptor().getMethods().get(72), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance()); @@ -80529,7 +78904,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(74), + getDescriptor().getMethods().get(73), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance()); @@ -80581,16 +78956,6 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MoveRegionResponse_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor; - private static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_hbase_pb_DispatchMergingRegionsRequest_fieldAccessorTable; - private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor; - private static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_hbase_pb_DispatchMergingRegionsResponse_fieldAccessorTable; - private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -81216,390 +79581,381 @@ public final class MasterProtos { "equest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Region", "Specifier\022.\n\020dest_server_name\030\002 \001(\0132\024.hb" + "ase.pb.ServerName\"\024\n\022MoveRegionResponse\"" + - "\274\001\n\035DispatchMergingRegionsRequest\022+\n\010reg" + - "ion_a\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\022+" + - "\n\010region_b\030\002 \002(\0132\031.hbase.pb.RegionSpecif" + - "ier\022\027\n\010forcible\030\003 \001(\010:\005false\022\026\n\013nonce_gr" + - "oup\030\004 \001(\004:\0010\022\020\n\005nonce\030\005 \001(\004:\0010\"1\n\036Dispat" + - "chMergingRegionsResponse\022\017\n\007proc_id\030\001 \001(" + - "\004\"\210\001\n\030MergeTableRegionsRequest\022)\n\006region" + - "\030\001 \003(\0132\031.hbase.pb.RegionSpecifier\022\027\n\010for", - "cible\030\003 \001(\010:\005false\022\026\n\013nonce_group\030\004 \001(\004:" + - "\0010\022\020\n\005nonce\030\005 \001(\004:\0010\",\n\031MergeTableRegion" + - "sResponse\022\017\n\007proc_id\030\001 \001(\004\"@\n\023AssignRegi" + - "onRequest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Reg" + - "ionSpecifier\"\026\n\024AssignRegionResponse\"X\n\025" + - "UnassignRegionRequest\022)\n\006region\030\001 \002(\0132\031." + - "hbase.pb.RegionSpecifier\022\024\n\005force\030\002 \001(\010:" + - "\005false\"\030\n\026UnassignRegionResponse\"A\n\024Offl" + - "ineRegionRequest\022)\n\006region\030\001 \002(\0132\031.hbase" + - ".pb.RegionSpecifier\"\027\n\025OfflineRegionResp", - "onse\"\201\001\n\027SplitTableRegionRequest\022)\n\013regi" + - "on_info\030\001 \002(\0132\024.hbase.pb.RegionInfo\022\021\n\ts" + - "plit_row\030\002 \001(\014\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020" + - "\n\005nonce\030\004 \001(\004:\0010\"+\n\030SplitTableRegionResp" + - "onse\022\017\n\007proc_id\030\001 \001(\004\"\177\n\022CreateTableRequ" + - "est\022+\n\014table_schema\030\001 \002(\0132\025.hbase.pb.Tab" + - "leSchema\022\022\n\nsplit_keys\030\002 \003(\014\022\026\n\013nonce_gr" + - "oup\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"&\n\023Create" + - "TableResponse\022\017\n\007proc_id\030\001 \001(\004\"g\n\022Delete" + - "TableRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase", - ".pb.TableName\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n" + - "\005nonce\030\003 \001(\004:\0010\"&\n\023DeleteTableResponse\022\017" + - "\n\007proc_id\030\001 \001(\004\"\207\001\n\024TruncateTableRequest" + - "\022&\n\ttableName\030\001 \002(\0132\023.hbase.pb.TableName" + - "\022\035\n\016preserveSplits\030\002 \001(\010:\005false\022\026\n\013nonce" + - "_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"(\n\025Tru" + - "ncateTableResponse\022\017\n\007proc_id\030\001 \001(\004\"g\n\022E" + - "nableTableRequest\022\'\n\ntable_name\030\001 \002(\0132\023." + - "hbase.pb.TableName\022\026\n\013nonce_group\030\002 \001(\004:" + - "\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"&\n\023EnableTableRespo", - "nse\022\017\n\007proc_id\030\001 \001(\004\"h\n\023DisableTableRequ" + - "est\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Table" + - "Name\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 " + - "\001(\004:\0010\"\'\n\024DisableTableResponse\022\017\n\007proc_i" + - "d\030\001 \001(\004\"\224\001\n\022ModifyTableRequest\022\'\n\ntable_" + - "name\030\001 \002(\0132\023.hbase.pb.TableName\022+\n\014table" + - "_schema\030\002 \002(\0132\025.hbase.pb.TableSchema\022\026\n\013" + - "nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"&" + - "\n\023ModifyTableResponse\022\017\n\007proc_id\030\001 \001(\004\"~" + - "\n\026CreateNamespaceRequest\022:\n\023namespaceDes", - "criptor\030\001 \002(\0132\035.hbase.pb.NamespaceDescri" + - "ptor\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 " + - "\001(\004:\0010\"*\n\027CreateNamespaceResponse\022\017\n\007pro" + - "c_id\030\001 \001(\004\"Y\n\026DeleteNamespaceRequest\022\025\n\r" + - "namespaceName\030\001 \002(\t\022\026\n\013nonce_group\030\002 \001(\004" + - ":\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"*\n\027DeleteNamespace" + - "Response\022\017\n\007proc_id\030\001 \001(\004\"~\n\026ModifyNames" + - "paceRequest\022:\n\023namespaceDescriptor\030\001 \002(\013" + - "2\035.hbase.pb.NamespaceDescriptor\022\026\n\013nonce" + - "_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"*\n\027Mod", - "ifyNamespaceResponse\022\017\n\007proc_id\030\001 \001(\004\"6\n" + - "\035GetNamespaceDescriptorRequest\022\025\n\rnamesp" + - "aceName\030\001 \002(\t\"\\\n\036GetNamespaceDescriptorR" + - "esponse\022:\n\023namespaceDescriptor\030\001 \002(\0132\035.h" + - "base.pb.NamespaceDescriptor\"!\n\037ListNames" + - "paceDescriptorsRequest\"^\n ListNamespaceD" + - "escriptorsResponse\022:\n\023namespaceDescripto" + - "r\030\001 \003(\0132\035.hbase.pb.NamespaceDescriptor\"?" + - "\n&ListTableDescriptorsByNamespaceRequest" + - "\022\025\n\rnamespaceName\030\001 \002(\t\"U\n\'ListTableDesc", - "riptorsByNamespaceResponse\022*\n\013tableSchem" + - "a\030\001 \003(\0132\025.hbase.pb.TableSchema\"9\n ListTa" + - "bleNamesByNamespaceRequest\022\025\n\rnamespaceN" + - "ame\030\001 \002(\t\"K\n!ListTableNamesByNamespaceRe" + - "sponse\022&\n\ttableName\030\001 \003(\0132\023.hbase.pb.Tab" + - "leName\"\021\n\017ShutdownRequest\"\022\n\020ShutdownRes" + - "ponse\"\023\n\021StopMasterRequest\"\024\n\022StopMaster" + - "Response\"\034\n\032IsInMaintenanceModeRequest\"8" + - "\n\033IsInMaintenanceModeResponse\022\031\n\021inMaint" + - "enanceMode\030\001 \002(\010\"\037\n\016BalanceRequest\022\r\n\005fo", - "rce\030\001 \001(\010\"\'\n\017BalanceResponse\022\024\n\014balancer" + - "_ran\030\001 \002(\010\"<\n\031SetBalancerRunningRequest\022" + - "\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010\"8\n\032SetB" + - "alancerRunningResponse\022\032\n\022prev_balance_v" + - "alue\030\001 \001(\010\"\032\n\030IsBalancerEnabledRequest\"," + - "\n\031IsBalancerEnabledResponse\022\017\n\007enabled\030\001" + - " \002(\010\"w\n\035SetSplitOrMergeEnabledRequest\022\017\n" + - "\007enabled\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010\0220\n\014s" + - "witch_types\030\003 \003(\0162\032.hbase.pb.MasterSwitc" + - "hType\"4\n\036SetSplitOrMergeEnabledResponse\022", - "\022\n\nprev_value\030\001 \003(\010\"O\n\034IsSplitOrMergeEna" + - "bledRequest\022/\n\013switch_type\030\001 \002(\0162\032.hbase" + - ".pb.MasterSwitchType\"0\n\035IsSplitOrMergeEn" + - "abledResponse\022\017\n\007enabled\030\001 \002(\010\"\022\n\020Normal" + - "izeRequest\"+\n\021NormalizeResponse\022\026\n\016norma" + - "lizer_ran\030\001 \002(\010\")\n\033SetNormalizerRunningR" + - "equest\022\n\n\002on\030\001 \002(\010\"=\n\034SetNormalizerRunni" + - "ngResponse\022\035\n\025prev_normalizer_value\030\001 \001(" + - "\010\"\034\n\032IsNormalizerEnabledRequest\".\n\033IsNor" + - "malizerEnabledResponse\022\017\n\007enabled\030\001 \002(\010\"", - "\027\n\025RunCatalogScanRequest\"-\n\026RunCatalogSc" + - "anResponse\022\023\n\013scan_result\030\001 \001(\005\"-\n\033Enabl" + - "eCatalogJanitorRequest\022\016\n\006enable\030\001 \002(\010\"2" + - "\n\034EnableCatalogJanitorResponse\022\022\n\nprev_v" + - "alue\030\001 \001(\010\" \n\036IsCatalogJanitorEnabledReq" + - "uest\"0\n\037IsCatalogJanitorEnabledResponse\022" + - "\r\n\005value\030\001 \002(\010\"\030\n\026RunCleanerChoreRequest" + - "\"4\n\027RunCleanerChoreResponse\022\031\n\021cleaner_c" + - "hore_ran\030\001 \002(\010\"+\n\035SetCleanerChoreRunning" + - "Request\022\n\n\002on\030\001 \002(\010\"4\n\036SetCleanerChoreRu", - "nningResponse\022\022\n\nprev_value\030\001 \001(\010\"\036\n\034IsC" + - "leanerChoreEnabledRequest\".\n\035IsCleanerCh" + - "oreEnabledResponse\022\r\n\005value\030\001 \002(\010\"B\n\017Sna" + - "pshotRequest\022/\n\010snapshot\030\001 \002(\0132\035.hbase.p" + - "b.SnapshotDescription\",\n\020SnapshotRespons" + - "e\022\030\n\020expected_timeout\030\001 \002(\003\"\036\n\034GetComple" + - "tedSnapshotsRequest\"Q\n\035GetCompletedSnaps" + - "hotsResponse\0220\n\tsnapshots\030\001 \003(\0132\035.hbase." + - "pb.SnapshotDescription\"H\n\025DeleteSnapshot" + - "Request\022/\n\010snapshot\030\001 \002(\0132\035.hbase.pb.Sna", - "pshotDescription\"\030\n\026DeleteSnapshotRespon" + - "se\"\216\001\n\026RestoreSnapshotRequest\022/\n\010snapsho" + - "t\030\001 \002(\0132\035.hbase.pb.SnapshotDescription\022\026" + - "\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010" + - "\022\031\n\nrestoreACL\030\004 \001(\010:\005false\"*\n\027RestoreSn" + - "apshotResponse\022\017\n\007proc_id\030\001 \002(\004\"H\n\025IsSna" + - "pshotDoneRequest\022/\n\010snapshot\030\001 \001(\0132\035.hba" + - "se.pb.SnapshotDescription\"^\n\026IsSnapshotD" + - "oneResponse\022\023\n\004done\030\001 \001(\010:\005false\022/\n\010snap" + - "shot\030\002 \001(\0132\035.hbase.pb.SnapshotDescriptio", - "n\"O\n\034IsRestoreSnapshotDoneRequest\022/\n\010sna" + - "pshot\030\001 \001(\0132\035.hbase.pb.SnapshotDescripti" + - "on\"4\n\035IsRestoreSnapshotDoneResponse\022\023\n\004d" + - "one\030\001 \001(\010:\005false\"F\n\033GetSchemaAlterStatus" + - "Request\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.T" + - "ableName\"T\n\034GetSchemaAlterStatusResponse" + - "\022\035\n\025yet_to_update_regions\030\001 \001(\r\022\025\n\rtotal" + - "_regions\030\002 \001(\r\"\213\001\n\032GetTableDescriptorsRe" + - "quest\022(\n\013table_names\030\001 \003(\0132\023.hbase.pb.Ta" + - "bleName\022\r\n\005regex\030\002 \001(\t\022!\n\022include_sys_ta", - "bles\030\003 \001(\010:\005false\022\021\n\tnamespace\030\004 \001(\t\"J\n\033" + - "GetTableDescriptorsResponse\022+\n\014table_sch" + - "ema\030\001 \003(\0132\025.hbase.pb.TableSchema\"[\n\024GetT" + - "ableNamesRequest\022\r\n\005regex\030\001 \001(\t\022!\n\022inclu" + - "de_sys_tables\030\002 \001(\010:\005false\022\021\n\tnamespace\030" + - "\003 \001(\t\"A\n\025GetTableNamesResponse\022(\n\013table_" + - "names\030\001 \003(\0132\023.hbase.pb.TableName\"?\n\024GetT" + - "ableStateRequest\022\'\n\ntable_name\030\001 \002(\0132\023.h" + - "base.pb.TableName\"B\n\025GetTableStateRespon" + - "se\022)\n\013table_state\030\001 \002(\0132\024.hbase.pb.Table", - "State\"\031\n\027GetClusterStatusRequest\"K\n\030GetC" + - "lusterStatusResponse\022/\n\016cluster_status\030\001" + - " \002(\0132\027.hbase.pb.ClusterStatus\"\030\n\026IsMaste" + - "rRunningRequest\"4\n\027IsMasterRunningRespon" + - "se\022\031\n\021is_master_running\030\001 \002(\010\"I\n\024ExecPro" + - "cedureRequest\0221\n\tprocedure\030\001 \002(\0132\036.hbase" + - ".pb.ProcedureDescription\"F\n\025ExecProcedur" + - "eResponse\022\030\n\020expected_timeout\030\001 \001(\003\022\023\n\013r" + - "eturn_data\030\002 \001(\014\"K\n\026IsProcedureDoneReque" + - "st\0221\n\tprocedure\030\001 \001(\0132\036.hbase.pb.Procedu", - "reDescription\"`\n\027IsProcedureDoneResponse" + - "\022\023\n\004done\030\001 \001(\010:\005false\0220\n\010snapshot\030\002 \001(\0132" + - "\036.hbase.pb.ProcedureDescription\",\n\031GetPr" + - "ocedureResultRequest\022\017\n\007proc_id\030\001 \002(\004\"\375\001" + - "\n\032GetProcedureResultResponse\0229\n\005state\030\001 " + - "\002(\0162*.hbase.pb.GetProcedureResultRespons" + - "e.State\022\026\n\016submitted_time\030\002 \001(\004\022\023\n\013last_" + - "update\030\003 \001(\004\022\016\n\006result\030\004 \001(\014\0224\n\texceptio" + - "n\030\005 \001(\0132!.hbase.pb.ForeignExceptionMessa" + - "ge\"1\n\005State\022\r\n\tNOT_FOUND\020\000\022\013\n\007RUNNING\020\001\022", - "\014\n\010FINISHED\020\002\"M\n\025AbortProcedureRequest\022\017" + - "\n\007proc_id\030\001 \002(\004\022#\n\025mayInterruptIfRunning" + - "\030\002 \001(\010:\004true\"6\n\026AbortProcedureResponse\022\034" + - "\n\024is_procedure_aborted\030\001 \002(\010\"\027\n\025ListProc" + - "eduresRequest\"@\n\026ListProceduresResponse\022" + - "&\n\tprocedure\030\001 \003(\0132\023.hbase.pb.Procedure\"" + - "\022\n\020ListLocksRequest\"5\n\021ListLocksResponse" + - "\022 \n\004lock\030\001 \003(\0132\022.hbase.pb.LockInfo\"\377\001\n\017S" + - "etQuotaRequest\022\021\n\tuser_name\030\001 \001(\t\022\022\n\nuse" + - "r_group\030\002 \001(\t\022\021\n\tnamespace\030\003 \001(\t\022\'\n\ntabl", - "e_name\030\004 \001(\0132\023.hbase.pb.TableName\022\022\n\nrem" + - "ove_all\030\005 \001(\010\022\026\n\016bypass_globals\030\006 \001(\010\022+\n" + - "\010throttle\030\007 \001(\0132\031.hbase.pb.ThrottleReque" + - "st\0220\n\013space_limit\030\010 \001(\0132\033.hbase.pb.Space" + - "LimitRequest\"\022\n\020SetQuotaResponse\"J\n\037Majo" + - "rCompactionTimestampRequest\022\'\n\ntable_nam" + - "e\030\001 \002(\0132\023.hbase.pb.TableName\"U\n(MajorCom" + - "pactionTimestampForRegionRequest\022)\n\006regi" + - "on\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\"@\n M" + - "ajorCompactionTimestampResponse\022\034\n\024compa", - "ction_timestamp\030\001 \002(\003\"\035\n\033SecurityCapabil" + - "itiesRequest\"\354\001\n\034SecurityCapabilitiesRes" + - "ponse\022G\n\014capabilities\030\001 \003(\01621.hbase.pb.S" + - "ecurityCapabilitiesResponse.Capability\"\202" + - "\001\n\nCapability\022\031\n\025SIMPLE_AUTHENTICATION\020\000" + - "\022\031\n\025SECURE_AUTHENTICATION\020\001\022\021\n\rAUTHORIZA" + - "TION\020\002\022\026\n\022CELL_AUTHORIZATION\020\003\022\023\n\017CELL_V" + - "ISIBILITY\020\004\"\"\n ListDrainingRegionServers" + - "Request\"N\n!ListDrainingRegionServersResp" + - "onse\022)\n\013server_name\030\001 \003(\0132\024.hbase.pb.Ser", - "verName\"F\n\031DrainRegionServersRequest\022)\n\013" + - "server_name\030\001 \003(\0132\024.hbase.pb.ServerName\"" + - "\034\n\032DrainRegionServersResponse\"P\n#RemoveD" + - "rainFromRegionServersRequest\022)\n\013server_n" + - "ame\030\001 \003(\0132\024.hbase.pb.ServerName\"&\n$Remov" + - "eDrainFromRegionServersResponse*(\n\020Maste" + - "rSwitchType\022\t\n\005SPLIT\020\000\022\t\n\005MERGE\020\0012\2227\n\rMa" + - "sterService\022e\n\024GetSchemaAlterStatus\022%.hb" + - "ase.pb.GetSchemaAlterStatusRequest\032&.hba" + - "se.pb.GetSchemaAlterStatusResponse\022b\n\023Ge", - "tTableDescriptors\022$.hbase.pb.GetTableDes" + - "criptorsRequest\032%.hbase.pb.GetTableDescr" + - "iptorsResponse\022P\n\rGetTableNames\022\036.hbase." + - "pb.GetTableNamesRequest\032\037.hbase.pb.GetTa" + - "bleNamesResponse\022Y\n\020GetClusterStatus\022!.h" + - "base.pb.GetClusterStatusRequest\032\".hbase." + - "pb.GetClusterStatusResponse\022V\n\017IsMasterR" + - "unning\022 .hbase.pb.IsMasterRunningRequest" + - "\032!.hbase.pb.IsMasterRunningResponse\022D\n\tA" + - "ddColumn\022\032.hbase.pb.AddColumnRequest\032\033.h", - "base.pb.AddColumnResponse\022M\n\014DeleteColum" + - "n\022\035.hbase.pb.DeleteColumnRequest\032\036.hbase" + - ".pb.DeleteColumnResponse\022M\n\014ModifyColumn" + - "\022\035.hbase.pb.ModifyColumnRequest\032\036.hbase." + - "pb.ModifyColumnResponse\022k\n\026DispatchMergi" + - "ngRegions\022\'.hbase.pb.DispatchMergingRegi" + - "onsRequest\032(.hbase.pb.DispatchMergingReg" + - "ionsResponse\022G\n\nMoveRegion\022\033.hbase.pb.Mo" + - "veRegionRequest\032\034.hbase.pb.MoveRegionRes" + - "ponse\022\\\n\021MergeTableRegions\022\".hbase.pb.Me", - "rgeTableRegionsRequest\032#.hbase.pb.MergeT" + - "ableRegionsResponse\022M\n\014AssignRegion\022\035.hb" + - "ase.pb.AssignRegionRequest\032\036.hbase.pb.As" + - "signRegionResponse\022S\n\016UnassignRegion\022\037.h" + - "base.pb.UnassignRegionRequest\032 .hbase.pb" + - ".UnassignRegionResponse\022P\n\rOfflineRegion" + - "\022\036.hbase.pb.OfflineRegionRequest\032\037.hbase" + - ".pb.OfflineRegionResponse\022T\n\013SplitRegion" + - "\022!.hbase.pb.SplitTableRegionRequest\032\".hb" + - "ase.pb.SplitTableRegionResponse\022J\n\013Delet", - "eTable\022\034.hbase.pb.DeleteTableRequest\032\035.h" + - "base.pb.DeleteTableResponse\022P\n\rtruncateT" + - "able\022\036.hbase.pb.TruncateTableRequest\032\037.h" + - "base.pb.TruncateTableResponse\022J\n\013EnableT" + - "able\022\034.hbase.pb.EnableTableRequest\032\035.hba" + - "se.pb.EnableTableResponse\022M\n\014DisableTabl" + - "e\022\035.hbase.pb.DisableTableRequest\032\036.hbase" + - ".pb.DisableTableResponse\022J\n\013ModifyTable\022" + - "\034.hbase.pb.ModifyTableRequest\032\035.hbase.pb" + - ".ModifyTableResponse\022J\n\013CreateTable\022\034.hb", - "ase.pb.CreateTableRequest\032\035.hbase.pb.Cre" + - "ateTableResponse\022A\n\010Shutdown\022\031.hbase.pb." + - "ShutdownRequest\032\032.hbase.pb.ShutdownRespo" + - "nse\022G\n\nStopMaster\022\033.hbase.pb.StopMasterR" + - "equest\032\034.hbase.pb.StopMasterResponse\022h\n\031" + - "IsMasterInMaintenanceMode\022$.hbase.pb.IsI" + - "nMaintenanceModeRequest\032%.hbase.pb.IsInM" + - "aintenanceModeResponse\022>\n\007Balance\022\030.hbas" + - "e.pb.BalanceRequest\032\031.hbase.pb.BalanceRe" + - "sponse\022_\n\022SetBalancerRunning\022#.hbase.pb.", - "SetBalancerRunningRequest\032$.hbase.pb.Set" + - "BalancerRunningResponse\022\\\n\021IsBalancerEna" + - "bled\022\".hbase.pb.IsBalancerEnabledRequest" + - "\032#.hbase.pb.IsBalancerEnabledResponse\022k\n" + - "\026SetSplitOrMergeEnabled\022\'.hbase.pb.SetSp" + - "litOrMergeEnabledRequest\032(.hbase.pb.SetS" + - "plitOrMergeEnabledResponse\022h\n\025IsSplitOrM" + - "ergeEnabled\022&.hbase.pb.IsSplitOrMergeEna" + - "bledRequest\032\'.hbase.pb.IsSplitOrMergeEna" + - "bledResponse\022D\n\tNormalize\022\032.hbase.pb.Nor", - "malizeRequest\032\033.hbase.pb.NormalizeRespon" + - "se\022e\n\024SetNormalizerRunning\022%.hbase.pb.Se" + - "tNormalizerRunningRequest\032&.hbase.pb.Set" + - "NormalizerRunningResponse\022b\n\023IsNormalize" + - "rEnabled\022$.hbase.pb.IsNormalizerEnabledR" + - "equest\032%.hbase.pb.IsNormalizerEnabledRes" + - "ponse\022S\n\016RunCatalogScan\022\037.hbase.pb.RunCa" + - "talogScanRequest\032 .hbase.pb.RunCatalogSc" + - "anResponse\022e\n\024EnableCatalogJanitor\022%.hba" + - "se.pb.EnableCatalogJanitorRequest\032&.hbas", - "e.pb.EnableCatalogJanitorResponse\022n\n\027IsC" + - "atalogJanitorEnabled\022(.hbase.pb.IsCatalo" + - "gJanitorEnabledRequest\032).hbase.pb.IsCata" + - "logJanitorEnabledResponse\022V\n\017RunCleanerC" + - "hore\022 .hbase.pb.RunCleanerChoreRequest\032!" + - ".hbase.pb.RunCleanerChoreResponse\022k\n\026Set" + - "CleanerChoreRunning\022\'.hbase.pb.SetCleane" + - "rChoreRunningRequest\032(.hbase.pb.SetClean" + - "erChoreRunningResponse\022h\n\025IsCleanerChore" + - "Enabled\022&.hbase.pb.IsCleanerChoreEnabled", - "Request\032\'.hbase.pb.IsCleanerChoreEnabled" + - "Response\022^\n\021ExecMasterService\022#.hbase.pb" + - ".CoprocessorServiceRequest\032$.hbase.pb.Co" + - "processorServiceResponse\022A\n\010Snapshot\022\031.h" + - "base.pb.SnapshotRequest\032\032.hbase.pb.Snaps" + - "hotResponse\022h\n\025GetCompletedSnapshots\022&.h" + - "base.pb.GetCompletedSnapshotsRequest\032\'.h" + - "base.pb.GetCompletedSnapshotsResponse\022S\n" + - "\016DeleteSnapshot\022\037.hbase.pb.DeleteSnapsho" + - "tRequest\032 .hbase.pb.DeleteSnapshotRespon", - "se\022S\n\016IsSnapshotDone\022\037.hbase.pb.IsSnapsh" + - "otDoneRequest\032 .hbase.pb.IsSnapshotDoneR" + - "esponse\022V\n\017RestoreSnapshot\022 .hbase.pb.Re" + - "storeSnapshotRequest\032!.hbase.pb.RestoreS" + - "napshotResponse\022P\n\rExecProcedure\022\036.hbase" + - ".pb.ExecProcedureRequest\032\037.hbase.pb.Exec" + - "ProcedureResponse\022W\n\024ExecProcedureWithRe" + - "t\022\036.hbase.pb.ExecProcedureRequest\032\037.hbas" + - "e.pb.ExecProcedureResponse\022V\n\017IsProcedur" + - "eDone\022 .hbase.pb.IsProcedureDoneRequest\032", - "!.hbase.pb.IsProcedureDoneResponse\022V\n\017Mo" + - "difyNamespace\022 .hbase.pb.ModifyNamespace" + - "Request\032!.hbase.pb.ModifyNamespaceRespon" + - "se\022V\n\017CreateNamespace\022 .hbase.pb.CreateN" + - "amespaceRequest\032!.hbase.pb.CreateNamespa" + - "ceResponse\022V\n\017DeleteNamespace\022 .hbase.pb" + - ".DeleteNamespaceRequest\032!.hbase.pb.Delet" + - "eNamespaceResponse\022k\n\026GetNamespaceDescri" + - "ptor\022\'.hbase.pb.GetNamespaceDescriptorRe" + - "quest\032(.hbase.pb.GetNamespaceDescriptorR", - "esponse\022q\n\030ListNamespaceDescriptors\022).hb" + - "ase.pb.ListNamespaceDescriptorsRequest\032*" + - ".hbase.pb.ListNamespaceDescriptorsRespon" + - "se\022\206\001\n\037ListTableDescriptorsByNamespace\0220" + - ".hbase.pb.ListTableDescriptorsByNamespac" + - "eRequest\0321.hbase.pb.ListTableDescriptors" + - "ByNamespaceResponse\022t\n\031ListTableNamesByN" + - "amespace\022*.hbase.pb.ListTableNamesByName" + - "spaceRequest\032+.hbase.pb.ListTableNamesBy" + - "NamespaceResponse\022P\n\rGetTableState\022\036.hba", - "se.pb.GetTableStateRequest\032\037.hbase.pb.Ge" + - "tTableStateResponse\022A\n\010SetQuota\022\031.hbase." + - "pb.SetQuotaRequest\032\032.hbase.pb.SetQuotaRe" + - "sponse\022x\n\037getLastMajorCompactionTimestam" + - "p\022).hbase.pb.MajorCompactionTimestampReq" + - "uest\032*.hbase.pb.MajorCompactionTimestamp" + - "Response\022\212\001\n(getLastMajorCompactionTimes" + - "tampForRegion\0222.hbase.pb.MajorCompaction" + - "TimestampForRegionRequest\032*.hbase.pb.Maj" + - "orCompactionTimestampResponse\022_\n\022getProc", - "edureResult\022#.hbase.pb.GetProcedureResul" + - "tRequest\032$.hbase.pb.GetProcedureResultRe" + - "sponse\022h\n\027getSecurityCapabilities\022%.hbas" + - "e.pb.SecurityCapabilitiesRequest\032&.hbase" + - ".pb.SecurityCapabilitiesResponse\022S\n\016Abor" + - "tProcedure\022\037.hbase.pb.AbortProcedureRequ" + - "est\032 .hbase.pb.AbortProcedureResponse\022S\n" + - "\016ListProcedures\022\037.hbase.pb.ListProcedure" + - "sRequest\032 .hbase.pb.ListProceduresRespon" + - "se\022D\n\tListLocks\022\032.hbase.pb.ListLocksRequ", - "est\032\033.hbase.pb.ListLocksResponse\022_\n\022AddR" + - "eplicationPeer\022#.hbase.pb.AddReplication" + - "PeerRequest\032$.hbase.pb.AddReplicationPee" + - "rResponse\022h\n\025RemoveReplicationPeer\022&.hba" + - "se.pb.RemoveReplicationPeerRequest\032\'.hba" + - "se.pb.RemoveReplicationPeerResponse\022h\n\025E" + - "nableReplicationPeer\022&.hbase.pb.EnableRe" + - "plicationPeerRequest\032\'.hbase.pb.EnableRe" + - "plicationPeerResponse\022k\n\026DisableReplicat" + - "ionPeer\022\'.hbase.pb.DisableReplicationPee", - "rRequest\032(.hbase.pb.DisableReplicationPe" + - "erResponse\022q\n\030GetReplicationPeerConfig\022)" + - ".hbase.pb.GetReplicationPeerConfigReques" + - "t\032*.hbase.pb.GetReplicationPeerConfigRes" + - "ponse\022z\n\033UpdateReplicationPeerConfig\022,.h" + - "base.pb.UpdateReplicationPeerConfigReque" + - "st\032-.hbase.pb.UpdateReplicationPeerConfi" + - "gResponse\022e\n\024ListReplicationPeers\022%.hbas" + - "e.pb.ListReplicationPeersRequest\032&.hbase" + - ".pb.ListReplicationPeersResponse\022t\n\031list", - "DrainingRegionServers\022*.hbase.pb.ListDra" + - "iningRegionServersRequest\032+.hbase.pb.Lis" + - "tDrainingRegionServersResponse\022_\n\022drainR" + - "egionServers\022#.hbase.pb.DrainRegionServe" + - "rsRequest\032$.hbase.pb.DrainRegionServersR" + - "esponse\022}\n\034removeDrainFromRegionServers\022" + - "-.hbase.pb.RemoveDrainFromRegionServersR" + - "equest\032..hbase.pb.RemoveDrainFromRegionS" + - "erversResponse\022q\n\030GetSpaceQuotaRegionSiz" + - "es\022).hbase.pb.GetSpaceQuotaRegionSizesRe", - "quest\032*.hbase.pb.GetSpaceQuotaRegionSize" + - "sResponse\022S\n\016GetQuotaStates\022\037.hbase.pb.G" + - "etQuotaStatesRequest\032 .hbase.pb.GetQuota" + - "StatesResponseBI\n1org.apache.hadoop.hbas" + - "e.shaded.protobuf.generatedB\014MasterProto" + - "sH\001\210\001\001\240\001\001" + "\210\001\n\030MergeTableRegionsRequest\022)\n\006region\030\001" + + " \003(\0132\031.hbase.pb.RegionSpecifier\022\027\n\010forci" + + "ble\030\003 \001(\010:\005false\022\026\n\013nonce_group\030\004 \001(\004:\0010" + + "\022\020\n\005nonce\030\005 \001(\004:\0010\",\n\031MergeTableRegionsR" + + "esponse\022\017\n\007proc_id\030\001 \001(\004\"@\n\023AssignRegion" + + "Request\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Regio" + + "nSpecifier\"\026\n\024AssignRegionResponse\"X\n\025Un" + + "assignRegionRequest\022)\n\006region\030\001 \002(\0132\031.hb", + "ase.pb.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005f" + + "alse\"\030\n\026UnassignRegionResponse\"A\n\024Offlin" + + "eRegionRequest\022)\n\006region\030\001 \002(\0132\031.hbase.p" + + "b.RegionSpecifier\"\027\n\025OfflineRegionRespon" + + "se\"\201\001\n\027SplitTableRegionRequest\022)\n\013region" + + "_info\030\001 \002(\0132\024.hbase.pb.RegionInfo\022\021\n\tspl" + + "it_row\030\002 \001(\014\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005" + + "nonce\030\004 \001(\004:\0010\"+\n\030SplitTableRegionRespon" + + "se\022\017\n\007proc_id\030\001 \001(\004\"\177\n\022CreateTableReques" + + "t\022+\n\014table_schema\030\001 \002(\0132\025.hbase.pb.Table", + "Schema\022\022\n\nsplit_keys\030\002 \003(\014\022\026\n\013nonce_grou" + + "p\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"&\n\023CreateTa" + + "bleResponse\022\017\n\007proc_id\030\001 \001(\004\"g\n\022DeleteTa" + + "bleRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.p" + + "b.TableName\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005n" + + "once\030\003 \001(\004:\0010\"&\n\023DeleteTableResponse\022\017\n\007" + + "proc_id\030\001 \001(\004\"\207\001\n\024TruncateTableRequest\022&" + + "\n\ttableName\030\001 \002(\0132\023.hbase.pb.TableName\022\035" + + "\n\016preserveSplits\030\002 \001(\010:\005false\022\026\n\013nonce_g" + + "roup\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"(\n\025Trunc", + "ateTableResponse\022\017\n\007proc_id\030\001 \001(\004\"g\n\022Ena" + + "bleTableRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hb" + + "ase.pb.TableName\022\026\n\013nonce_group\030\002 \001(\004:\0010" + + "\022\020\n\005nonce\030\003 \001(\004:\0010\"&\n\023EnableTableRespons" + + "e\022\017\n\007proc_id\030\001 \001(\004\"h\n\023DisableTableReques" + + "t\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.TableNa" + + "me\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(" + + "\004:\0010\"\'\n\024DisableTableResponse\022\017\n\007proc_id\030" + + "\001 \001(\004\"\224\001\n\022ModifyTableRequest\022\'\n\ntable_na" + + "me\030\001 \002(\0132\023.hbase.pb.TableName\022+\n\014table_s", + "chema\030\002 \002(\0132\025.hbase.pb.TableSchema\022\026\n\013no" + + "nce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"&\n\023" + + "ModifyTableResponse\022\017\n\007proc_id\030\001 \001(\004\"~\n\026" + + "CreateNamespaceRequest\022:\n\023namespaceDescr" + + "iptor\030\001 \002(\0132\035.hbase.pb.NamespaceDescript" + + "or\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(" + + "\004:\0010\"*\n\027CreateNamespaceResponse\022\017\n\007proc_" + + "id\030\001 \001(\004\"Y\n\026DeleteNamespaceRequest\022\025\n\rna" + + "mespaceName\030\001 \002(\t\022\026\n\013nonce_group\030\002 \001(\004:\001" + + "0\022\020\n\005nonce\030\003 \001(\004:\0010\"*\n\027DeleteNamespaceRe", + "sponse\022\017\n\007proc_id\030\001 \001(\004\"~\n\026ModifyNamespa" + + "ceRequest\022:\n\023namespaceDescriptor\030\001 \002(\0132\035" + + ".hbase.pb.NamespaceDescriptor\022\026\n\013nonce_g" + + "roup\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"*\n\027Modif" + + "yNamespaceResponse\022\017\n\007proc_id\030\001 \001(\004\"6\n\035G" + + "etNamespaceDescriptorRequest\022\025\n\rnamespac" + + "eName\030\001 \002(\t\"\\\n\036GetNamespaceDescriptorRes" + + "ponse\022:\n\023namespaceDescriptor\030\001 \002(\0132\035.hba" + + "se.pb.NamespaceDescriptor\"!\n\037ListNamespa" + + "ceDescriptorsRequest\"^\n ListNamespaceDes", + "criptorsResponse\022:\n\023namespaceDescriptor\030" + + "\001 \003(\0132\035.hbase.pb.NamespaceDescriptor\"?\n&" + + "ListTableDescriptorsByNamespaceRequest\022\025" + + "\n\rnamespaceName\030\001 \002(\t\"U\n\'ListTableDescri" + + "ptorsByNamespaceResponse\022*\n\013tableSchema\030" + + "\001 \003(\0132\025.hbase.pb.TableSchema\"9\n ListTabl" + + "eNamesByNamespaceRequest\022\025\n\rnamespaceNam" + + "e\030\001 \002(\t\"K\n!ListTableNamesByNamespaceResp" + + "onse\022&\n\ttableName\030\001 \003(\0132\023.hbase.pb.Table" + + "Name\"\021\n\017ShutdownRequest\"\022\n\020ShutdownRespo", + "nse\"\023\n\021StopMasterRequest\"\024\n\022StopMasterRe" + + "sponse\"\034\n\032IsInMaintenanceModeRequest\"8\n\033" + + "IsInMaintenanceModeResponse\022\031\n\021inMainten" + + "anceMode\030\001 \002(\010\"\037\n\016BalanceRequest\022\r\n\005forc" + + "e\030\001 \001(\010\"\'\n\017BalanceResponse\022\024\n\014balancer_r" + + "an\030\001 \002(\010\"<\n\031SetBalancerRunningRequest\022\n\n" + + "\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010\"8\n\032SetBal" + + "ancerRunningResponse\022\032\n\022prev_balance_val" + + "ue\030\001 \001(\010\"\032\n\030IsBalancerEnabledRequest\",\n\031" + + "IsBalancerEnabledResponse\022\017\n\007enabled\030\001 \002", + "(\010\"w\n\035SetSplitOrMergeEnabledRequest\022\017\n\007e" + + "nabled\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010\0220\n\014swi" + + "tch_types\030\003 \003(\0162\032.hbase.pb.MasterSwitchT" + + "ype\"4\n\036SetSplitOrMergeEnabledResponse\022\022\n" + + "\nprev_value\030\001 \003(\010\"O\n\034IsSplitOrMergeEnabl" + + "edRequest\022/\n\013switch_type\030\001 \002(\0162\032.hbase.p" + + "b.MasterSwitchType\"0\n\035IsSplitOrMergeEnab" + + "ledResponse\022\017\n\007enabled\030\001 \002(\010\"\022\n\020Normaliz" + + "eRequest\"+\n\021NormalizeResponse\022\026\n\016normali" + + "zer_ran\030\001 \002(\010\")\n\033SetNormalizerRunningReq", + "uest\022\n\n\002on\030\001 \002(\010\"=\n\034SetNormalizerRunning" + + "Response\022\035\n\025prev_normalizer_value\030\001 \001(\010\"" + + "\034\n\032IsNormalizerEnabledRequest\".\n\033IsNorma" + + "lizerEnabledResponse\022\017\n\007enabled\030\001 \002(\010\"\027\n" + + "\025RunCatalogScanRequest\"-\n\026RunCatalogScan" + + "Response\022\023\n\013scan_result\030\001 \001(\005\"-\n\033EnableC" + + "atalogJanitorRequest\022\016\n\006enable\030\001 \002(\010\"2\n\034" + + "EnableCatalogJanitorResponse\022\022\n\nprev_val" + + "ue\030\001 \001(\010\" \n\036IsCatalogJanitorEnabledReque" + + "st\"0\n\037IsCatalogJanitorEnabledResponse\022\r\n", + "\005value\030\001 \002(\010\"\030\n\026RunCleanerChoreRequest\"4" + + "\n\027RunCleanerChoreResponse\022\031\n\021cleaner_cho" + + "re_ran\030\001 \002(\010\"+\n\035SetCleanerChoreRunningRe" + + "quest\022\n\n\002on\030\001 \002(\010\"4\n\036SetCleanerChoreRunn" + + "ingResponse\022\022\n\nprev_value\030\001 \001(\010\"\036\n\034IsCle" + + "anerChoreEnabledRequest\".\n\035IsCleanerChor" + + "eEnabledResponse\022\r\n\005value\030\001 \002(\010\"B\n\017Snaps" + + "hotRequest\022/\n\010snapshot\030\001 \002(\0132\035.hbase.pb." + + "SnapshotDescription\",\n\020SnapshotResponse\022" + + "\030\n\020expected_timeout\030\001 \002(\003\"\036\n\034GetComplete", + "dSnapshotsRequest\"Q\n\035GetCompletedSnapsho" + + "tsResponse\0220\n\tsnapshots\030\001 \003(\0132\035.hbase.pb" + + ".SnapshotDescription\"H\n\025DeleteSnapshotRe" + + "quest\022/\n\010snapshot\030\001 \002(\0132\035.hbase.pb.Snaps" + + "hotDescription\"\030\n\026DeleteSnapshotResponse" + + "\"\216\001\n\026RestoreSnapshotRequest\022/\n\010snapshot\030" + + "\001 \002(\0132\035.hbase.pb.SnapshotDescription\022\026\n\013" + + "nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\022\031" + + "\n\nrestoreACL\030\004 \001(\010:\005false\"*\n\027RestoreSnap" + + "shotResponse\022\017\n\007proc_id\030\001 \002(\004\"H\n\025IsSnaps", + "hotDoneRequest\022/\n\010snapshot\030\001 \001(\0132\035.hbase" + + ".pb.SnapshotDescription\"^\n\026IsSnapshotDon" + + "eResponse\022\023\n\004done\030\001 \001(\010:\005false\022/\n\010snapsh" + + "ot\030\002 \001(\0132\035.hbase.pb.SnapshotDescription\"" + + "O\n\034IsRestoreSnapshotDoneRequest\022/\n\010snaps" + + "hot\030\001 \001(\0132\035.hbase.pb.SnapshotDescription" + + "\"4\n\035IsRestoreSnapshotDoneResponse\022\023\n\004don" + + "e\030\001 \001(\010:\005false\"F\n\033GetSchemaAlterStatusRe" + + "quest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Tab" + + "leName\"T\n\034GetSchemaAlterStatusResponse\022\035", + "\n\025yet_to_update_regions\030\001 \001(\r\022\025\n\rtotal_r" + + "egions\030\002 \001(\r\"\213\001\n\032GetTableDescriptorsRequ" + + "est\022(\n\013table_names\030\001 \003(\0132\023.hbase.pb.Tabl" + + "eName\022\r\n\005regex\030\002 \001(\t\022!\n\022include_sys_tabl" + + "es\030\003 \001(\010:\005false\022\021\n\tnamespace\030\004 \001(\t\"J\n\033Ge" + + "tTableDescriptorsResponse\022+\n\014table_schem" + + "a\030\001 \003(\0132\025.hbase.pb.TableSchema\"[\n\024GetTab" + + "leNamesRequest\022\r\n\005regex\030\001 \001(\t\022!\n\022include" + + "_sys_tables\030\002 \001(\010:\005false\022\021\n\tnamespace\030\003 " + + "\001(\t\"A\n\025GetTableNamesResponse\022(\n\013table_na", + "mes\030\001 \003(\0132\023.hbase.pb.TableName\"?\n\024GetTab" + + "leStateRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hba" + + "se.pb.TableName\"B\n\025GetTableStateResponse" + + "\022)\n\013table_state\030\001 \002(\0132\024.hbase.pb.TableSt" + + "ate\"\031\n\027GetClusterStatusRequest\"K\n\030GetClu" + + "sterStatusResponse\022/\n\016cluster_status\030\001 \002" + + "(\0132\027.hbase.pb.ClusterStatus\"\030\n\026IsMasterR" + + "unningRequest\"4\n\027IsMasterRunningResponse" + + "\022\031\n\021is_master_running\030\001 \002(\010\"I\n\024ExecProce" + + "dureRequest\0221\n\tprocedure\030\001 \002(\0132\036.hbase.p", + "b.ProcedureDescription\"F\n\025ExecProcedureR" + + "esponse\022\030\n\020expected_timeout\030\001 \001(\003\022\023\n\013ret" + + "urn_data\030\002 \001(\014\"K\n\026IsProcedureDoneRequest" + + "\0221\n\tprocedure\030\001 \001(\0132\036.hbase.pb.Procedure" + + "Description\"`\n\027IsProcedureDoneResponse\022\023" + + "\n\004done\030\001 \001(\010:\005false\0220\n\010snapshot\030\002 \001(\0132\036." + + "hbase.pb.ProcedureDescription\",\n\031GetProc" + + "edureResultRequest\022\017\n\007proc_id\030\001 \002(\004\"\375\001\n\032" + + "GetProcedureResultResponse\0229\n\005state\030\001 \002(" + + "\0162*.hbase.pb.GetProcedureResultResponse.", + "State\022\026\n\016submitted_time\030\002 \001(\004\022\023\n\013last_up" + + "date\030\003 \001(\004\022\016\n\006result\030\004 \001(\014\0224\n\texception\030" + + "\005 \001(\0132!.hbase.pb.ForeignExceptionMessage" + + "\"1\n\005State\022\r\n\tNOT_FOUND\020\000\022\013\n\007RUNNING\020\001\022\014\n" + + "\010FINISHED\020\002\"M\n\025AbortProcedureRequest\022\017\n\007" + + "proc_id\030\001 \002(\004\022#\n\025mayInterruptIfRunning\030\002" + + " \001(\010:\004true\"6\n\026AbortProcedureResponse\022\034\n\024" + + "is_procedure_aborted\030\001 \002(\010\"\027\n\025ListProced" + + "uresRequest\"@\n\026ListProceduresResponse\022&\n" + + "\tprocedure\030\001 \003(\0132\023.hbase.pb.Procedure\"\022\n", + "\020ListLocksRequest\"5\n\021ListLocksResponse\022 " + + "\n\004lock\030\001 \003(\0132\022.hbase.pb.LockInfo\"\377\001\n\017Set" + + "QuotaRequest\022\021\n\tuser_name\030\001 \001(\t\022\022\n\nuser_" + + "group\030\002 \001(\t\022\021\n\tnamespace\030\003 \001(\t\022\'\n\ntable_" + + "name\030\004 \001(\0132\023.hbase.pb.TableName\022\022\n\nremov" + + "e_all\030\005 \001(\010\022\026\n\016bypass_globals\030\006 \001(\010\022+\n\010t" + + "hrottle\030\007 \001(\0132\031.hbase.pb.ThrottleRequest" + + "\0220\n\013space_limit\030\010 \001(\0132\033.hbase.pb.SpaceLi" + + "mitRequest\"\022\n\020SetQuotaResponse\"J\n\037MajorC" + + "ompactionTimestampRequest\022\'\n\ntable_name\030", + "\001 \002(\0132\023.hbase.pb.TableName\"U\n(MajorCompa" + + "ctionTimestampForRegionRequest\022)\n\006region" + + "\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\"@\n Maj" + + "orCompactionTimestampResponse\022\034\n\024compact" + + "ion_timestamp\030\001 \002(\003\"\035\n\033SecurityCapabilit" + + "iesRequest\"\354\001\n\034SecurityCapabilitiesRespo" + + "nse\022G\n\014capabilities\030\001 \003(\01621.hbase.pb.Sec" + + "urityCapabilitiesResponse.Capability\"\202\001\n" + + "\nCapability\022\031\n\025SIMPLE_AUTHENTICATION\020\000\022\031" + + "\n\025SECURE_AUTHENTICATION\020\001\022\021\n\rAUTHORIZATI", + "ON\020\002\022\026\n\022CELL_AUTHORIZATION\020\003\022\023\n\017CELL_VIS" + + "IBILITY\020\004\"\"\n ListDrainingRegionServersRe" + + "quest\"N\n!ListDrainingRegionServersRespon" + + "se\022)\n\013server_name\030\001 \003(\0132\024.hbase.pb.Serve" + + "rName\"F\n\031DrainRegionServersRequest\022)\n\013se" + + "rver_name\030\001 \003(\0132\024.hbase.pb.ServerName\"\034\n" + + "\032DrainRegionServersResponse\"P\n#RemoveDra" + + "inFromRegionServersRequest\022)\n\013server_nam" + + "e\030\001 \003(\0132\024.hbase.pb.ServerName\"&\n$RemoveD" + + "rainFromRegionServersResponse*(\n\020MasterS", + "witchType\022\t\n\005SPLIT\020\000\022\t\n\005MERGE\020\0012\2456\n\rMast" + + "erService\022e\n\024GetSchemaAlterStatus\022%.hbas" + + "e.pb.GetSchemaAlterStatusRequest\032&.hbase" + + ".pb.GetSchemaAlterStatusResponse\022b\n\023GetT" + + "ableDescriptors\022$.hbase.pb.GetTableDescr" + + "iptorsRequest\032%.hbase.pb.GetTableDescrip" + + "torsResponse\022P\n\rGetTableNames\022\036.hbase.pb" + + ".GetTableNamesRequest\032\037.hbase.pb.GetTabl" + + "eNamesResponse\022Y\n\020GetClusterStatus\022!.hba" + + "se.pb.GetClusterStatusRequest\032\".hbase.pb", + ".GetClusterStatusResponse\022V\n\017IsMasterRun" + + "ning\022 .hbase.pb.IsMasterRunningRequest\032!" + + ".hbase.pb.IsMasterRunningResponse\022D\n\tAdd" + + "Column\022\032.hbase.pb.AddColumnRequest\032\033.hba" + + "se.pb.AddColumnResponse\022M\n\014DeleteColumn\022" + + "\035.hbase.pb.DeleteColumnRequest\032\036.hbase.p" + + "b.DeleteColumnResponse\022M\n\014ModifyColumn\022\035" + + ".hbase.pb.ModifyColumnRequest\032\036.hbase.pb" + + ".ModifyColumnResponse\022G\n\nMoveRegion\022\033.hb" + + "ase.pb.MoveRegionRequest\032\034.hbase.pb.Move", + "RegionResponse\022\\\n\021MergeTableRegions\022\".hb" + + "ase.pb.MergeTableRegionsRequest\032#.hbase." + + "pb.MergeTableRegionsResponse\022M\n\014AssignRe" + + "gion\022\035.hbase.pb.AssignRegionRequest\032\036.hb" + + "ase.pb.AssignRegionResponse\022S\n\016UnassignR" + + "egion\022\037.hbase.pb.UnassignRegionRequest\032 " + + ".hbase.pb.UnassignRegionResponse\022P\n\rOffl" + + "ineRegion\022\036.hbase.pb.OfflineRegionReques" + + "t\032\037.hbase.pb.OfflineRegionResponse\022T\n\013Sp" + + "litRegion\022!.hbase.pb.SplitTableRegionReq", + "uest\032\".hbase.pb.SplitTableRegionResponse" + + "\022J\n\013DeleteTable\022\034.hbase.pb.DeleteTableRe" + + "quest\032\035.hbase.pb.DeleteTableResponse\022P\n\r" + + "truncateTable\022\036.hbase.pb.TruncateTableRe" + + "quest\032\037.hbase.pb.TruncateTableResponse\022J" + + "\n\013EnableTable\022\034.hbase.pb.EnableTableRequ" + + "est\032\035.hbase.pb.EnableTableResponse\022M\n\014Di" + + "sableTable\022\035.hbase.pb.DisableTableReques" + + "t\032\036.hbase.pb.DisableTableResponse\022J\n\013Mod" + + "ifyTable\022\034.hbase.pb.ModifyTableRequest\032\035", + ".hbase.pb.ModifyTableResponse\022J\n\013CreateT" + + "able\022\034.hbase.pb.CreateTableRequest\032\035.hba" + + "se.pb.CreateTableResponse\022A\n\010Shutdown\022\031." + + "hbase.pb.ShutdownRequest\032\032.hbase.pb.Shut" + + "downResponse\022G\n\nStopMaster\022\033.hbase.pb.St" + + "opMasterRequest\032\034.hbase.pb.StopMasterRes" + + "ponse\022h\n\031IsMasterInMaintenanceMode\022$.hba" + + "se.pb.IsInMaintenanceModeRequest\032%.hbase" + + ".pb.IsInMaintenanceModeResponse\022>\n\007Balan" + + "ce\022\030.hbase.pb.BalanceRequest\032\031.hbase.pb.", + "BalanceResponse\022_\n\022SetBalancerRunning\022#." + + "hbase.pb.SetBalancerRunningRequest\032$.hba" + + "se.pb.SetBalancerRunningResponse\022\\\n\021IsBa" + + "lancerEnabled\022\".hbase.pb.IsBalancerEnabl" + + "edRequest\032#.hbase.pb.IsBalancerEnabledRe" + + "sponse\022k\n\026SetSplitOrMergeEnabled\022\'.hbase" + + ".pb.SetSplitOrMergeEnabledRequest\032(.hbas" + + "e.pb.SetSplitOrMergeEnabledResponse\022h\n\025I" + + "sSplitOrMergeEnabled\022&.hbase.pb.IsSplitO" + + "rMergeEnabledRequest\032\'.hbase.pb.IsSplitO", + "rMergeEnabledResponse\022D\n\tNormalize\022\032.hba" + + "se.pb.NormalizeRequest\032\033.hbase.pb.Normal" + + "izeResponse\022e\n\024SetNormalizerRunning\022%.hb" + + "ase.pb.SetNormalizerRunningRequest\032&.hba" + + "se.pb.SetNormalizerRunningResponse\022b\n\023Is" + + "NormalizerEnabled\022$.hbase.pb.IsNormalize" + + "rEnabledRequest\032%.hbase.pb.IsNormalizerE" + + "nabledResponse\022S\n\016RunCatalogScan\022\037.hbase" + + ".pb.RunCatalogScanRequest\032 .hbase.pb.Run" + + "CatalogScanResponse\022e\n\024EnableCatalogJani", + "tor\022%.hbase.pb.EnableCatalogJanitorReque" + + "st\032&.hbase.pb.EnableCatalogJanitorRespon" + + "se\022n\n\027IsCatalogJanitorEnabled\022(.hbase.pb" + + ".IsCatalogJanitorEnabledRequest\032).hbase." + + "pb.IsCatalogJanitorEnabledResponse\022V\n\017Ru" + + "nCleanerChore\022 .hbase.pb.RunCleanerChore" + + "Request\032!.hbase.pb.RunCleanerChoreRespon" + + "se\022k\n\026SetCleanerChoreRunning\022\'.hbase.pb." + + "SetCleanerChoreRunningRequest\032(.hbase.pb" + + ".SetCleanerChoreRunningResponse\022h\n\025IsCle", + "anerChoreEnabled\022&.hbase.pb.IsCleanerCho" + + "reEnabledRequest\032\'.hbase.pb.IsCleanerCho" + + "reEnabledResponse\022^\n\021ExecMasterService\022#" + + ".hbase.pb.CoprocessorServiceRequest\032$.hb" + + "ase.pb.CoprocessorServiceResponse\022A\n\010Sna" + + "pshot\022\031.hbase.pb.SnapshotRequest\032\032.hbase" + + ".pb.SnapshotResponse\022h\n\025GetCompletedSnap" + + "shots\022&.hbase.pb.GetCompletedSnapshotsRe" + + "quest\032\'.hbase.pb.GetCompletedSnapshotsRe" + + "sponse\022S\n\016DeleteSnapshot\022\037.hbase.pb.Dele", + "teSnapshotRequest\032 .hbase.pb.DeleteSnaps" + + "hotResponse\022S\n\016IsSnapshotDone\022\037.hbase.pb" + + ".IsSnapshotDoneRequest\032 .hbase.pb.IsSnap" + + "shotDoneResponse\022V\n\017RestoreSnapshot\022 .hb" + + "ase.pb.RestoreSnapshotRequest\032!.hbase.pb" + + ".RestoreSnapshotResponse\022P\n\rExecProcedur" + + "e\022\036.hbase.pb.ExecProcedureRequest\032\037.hbas" + + "e.pb.ExecProcedureResponse\022W\n\024ExecProced" + + "ureWithRet\022\036.hbase.pb.ExecProcedureReque" + + "st\032\037.hbase.pb.ExecProcedureResponse\022V\n\017I", + "sProcedureDone\022 .hbase.pb.IsProcedureDon" + + "eRequest\032!.hbase.pb.IsProcedureDoneRespo" + + "nse\022V\n\017ModifyNamespace\022 .hbase.pb.Modify" + + "NamespaceRequest\032!.hbase.pb.ModifyNamesp" + + "aceResponse\022V\n\017CreateNamespace\022 .hbase.p" + + "b.CreateNamespaceRequest\032!.hbase.pb.Crea" + + "teNamespaceResponse\022V\n\017DeleteNamespace\022 " + + ".hbase.pb.DeleteNamespaceRequest\032!.hbase" + + ".pb.DeleteNamespaceResponse\022k\n\026GetNamesp" + + "aceDescriptor\022\'.hbase.pb.GetNamespaceDes", + "criptorRequest\032(.hbase.pb.GetNamespaceDe" + + "scriptorResponse\022q\n\030ListNamespaceDescrip" + + "tors\022).hbase.pb.ListNamespaceDescriptors" + + "Request\032*.hbase.pb.ListNamespaceDescript" + + "orsResponse\022\206\001\n\037ListTableDescriptorsByNa" + + "mespace\0220.hbase.pb.ListTableDescriptorsB" + + "yNamespaceRequest\0321.hbase.pb.ListTableDe" + + "scriptorsByNamespaceResponse\022t\n\031ListTabl" + + "eNamesByNamespace\022*.hbase.pb.ListTableNa" + + "mesByNamespaceRequest\032+.hbase.pb.ListTab", + "leNamesByNamespaceResponse\022P\n\rGetTableSt" + + "ate\022\036.hbase.pb.GetTableStateRequest\032\037.hb" + + "ase.pb.GetTableStateResponse\022A\n\010SetQuota" + + "\022\031.hbase.pb.SetQuotaRequest\032\032.hbase.pb.S" + + "etQuotaResponse\022x\n\037getLastMajorCompactio" + + "nTimestamp\022).hbase.pb.MajorCompactionTim" + + "estampRequest\032*.hbase.pb.MajorCompaction" + + "TimestampResponse\022\212\001\n(getLastMajorCompac" + + "tionTimestampForRegion\0222.hbase.pb.MajorC" + + "ompactionTimestampForRegionRequest\032*.hba", + "se.pb.MajorCompactionTimestampResponse\022_" + + "\n\022getProcedureResult\022#.hbase.pb.GetProce" + + "dureResultRequest\032$.hbase.pb.GetProcedur" + + "eResultResponse\022h\n\027getSecurityCapabiliti" + + "es\022%.hbase.pb.SecurityCapabilitiesReques" + + "t\032&.hbase.pb.SecurityCapabilitiesRespons" + + "e\022S\n\016AbortProcedure\022\037.hbase.pb.AbortProc" + + "edureRequest\032 .hbase.pb.AbortProcedureRe" + + "sponse\022S\n\016ListProcedures\022\037.hbase.pb.List" + + "ProceduresRequest\032 .hbase.pb.ListProcedu", + "resResponse\022D\n\tListLocks\022\032.hbase.pb.List" + + "LocksRequest\032\033.hbase.pb.ListLocksRespons" + + "e\022_\n\022AddReplicationPeer\022#.hbase.pb.AddRe" + + "plicationPeerRequest\032$.hbase.pb.AddRepli" + + "cationPeerResponse\022h\n\025RemoveReplicationP" + + "eer\022&.hbase.pb.RemoveReplicationPeerRequ" + + "est\032\'.hbase.pb.RemoveReplicationPeerResp" + + "onse\022h\n\025EnableReplicationPeer\022&.hbase.pb" + + ".EnableReplicationPeerRequest\032\'.hbase.pb" + + ".EnableReplicationPeerResponse\022k\n\026Disabl", + "eReplicationPeer\022\'.hbase.pb.DisableRepli" + + "cationPeerRequest\032(.hbase.pb.DisableRepl" + + "icationPeerResponse\022q\n\030GetReplicationPee" + + "rConfig\022).hbase.pb.GetReplicationPeerCon" + + "figRequest\032*.hbase.pb.GetReplicationPeer" + + "ConfigResponse\022z\n\033UpdateReplicationPeerC" + + "onfig\022,.hbase.pb.UpdateReplicationPeerCo" + + "nfigRequest\032-.hbase.pb.UpdateReplication" + + "PeerConfigResponse\022e\n\024ListReplicationPee" + + "rs\022%.hbase.pb.ListReplicationPeersReques", + "t\032&.hbase.pb.ListReplicationPeersRespons" + + "e\022t\n\031listDrainingRegionServers\022*.hbase.p" + + "b.ListDrainingRegionServersRequest\032+.hba" + + "se.pb.ListDrainingRegionServersResponse\022" + + "_\n\022drainRegionServers\022#.hbase.pb.DrainRe" + + "gionServersRequest\032$.hbase.pb.DrainRegio" + + "nServersResponse\022}\n\034removeDrainFromRegio" + + "nServers\022-.hbase.pb.RemoveDrainFromRegio" + + "nServersRequest\032..hbase.pb.RemoveDrainFr" + + "omRegionServersResponse\022q\n\030GetSpaceQuota", + "RegionSizes\022).hbase.pb.GetSpaceQuotaRegi" + + "onSizesRequest\032*.hbase.pb.GetSpaceQuotaR" + + "egionSizesResponse\022S\n\016GetQuotaStates\022\037.h" + + "base.pb.GetQuotaStatesRequest\032 .hbase.pb" + + ".GetQuotaStatesResponseBI\n1org.apache.ha" + + "doop.hbase.shaded.protobuf.generatedB\014Ma" + + "sterProtosH\001\210\001\001\240\001\001" }; org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() { @@ -81670,728 +80026,716 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MoveRegionResponse_descriptor, new java.lang.String[] { }); - internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor = - getDescriptor().getMessageTypes().get(8); - internal_static_hbase_pb_DispatchMergingRegionsRequest_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_hbase_pb_DispatchMergingRegionsRequest_descriptor, - new java.lang.String[] { "RegionA", "RegionB", "Forcible", "NonceGroup", "Nonce", }); - internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor = - getDescriptor().getMessageTypes().get(9); - internal_static_hbase_pb_DispatchMergingRegionsResponse_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_hbase_pb_DispatchMergingRegionsResponse_descriptor, - new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_MergeTableRegionsRequest_descriptor = - getDescriptor().getMessageTypes().get(10); + getDescriptor().getMessageTypes().get(8); internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MergeTableRegionsRequest_descriptor, new java.lang.String[] { "Region", "Forcible", "NonceGroup", "Nonce", }); internal_static_hbase_pb_MergeTableRegionsResponse_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(9); internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MergeTableRegionsResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_AssignRegionRequest_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(10); internal_static_hbase_pb_AssignRegionRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AssignRegionRequest_descriptor, new java.lang.String[] { "Region", }); internal_static_hbase_pb_AssignRegionResponse_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(11); internal_static_hbase_pb_AssignRegionResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AssignRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_UnassignRegionRequest_descriptor = - getDescriptor().getMessageTypes().get(14); + getDescriptor().getMessageTypes().get(12); internal_static_hbase_pb_UnassignRegionRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UnassignRegionRequest_descriptor, new java.lang.String[] { "Region", "Force", }); internal_static_hbase_pb_UnassignRegionResponse_descriptor = - getDescriptor().getMessageTypes().get(15); + getDescriptor().getMessageTypes().get(13); internal_static_hbase_pb_UnassignRegionResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UnassignRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_OfflineRegionRequest_descriptor = - getDescriptor().getMessageTypes().get(16); + getDescriptor().getMessageTypes().get(14); internal_static_hbase_pb_OfflineRegionRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_OfflineRegionRequest_descriptor, new java.lang.String[] { "Region", }); internal_static_hbase_pb_OfflineRegionResponse_descriptor = - getDescriptor().getMessageTypes().get(17); + getDescriptor().getMessageTypes().get(15); internal_static_hbase_pb_OfflineRegionResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_OfflineRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_SplitTableRegionRequest_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(16); internal_static_hbase_pb_SplitTableRegionRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SplitTableRegionRequest_descriptor, new java.lang.String[] { "RegionInfo", "SplitRow", "NonceGroup", "Nonce", }); internal_static_hbase_pb_SplitTableRegionResponse_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(17); internal_static_hbase_pb_SplitTableRegionResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SplitTableRegionResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_CreateTableRequest_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(18); internal_static_hbase_pb_CreateTableRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateTableRequest_descriptor, new java.lang.String[] { "TableSchema", "SplitKeys", "NonceGroup", "Nonce", }); internal_static_hbase_pb_CreateTableResponse_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(19); internal_static_hbase_pb_CreateTableResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DeleteTableRequest_descriptor = - getDescriptor().getMessageTypes().get(22); + getDescriptor().getMessageTypes().get(20); internal_static_hbase_pb_DeleteTableRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteTableRequest_descriptor, new java.lang.String[] { "TableName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DeleteTableResponse_descriptor = - getDescriptor().getMessageTypes().get(23); + getDescriptor().getMessageTypes().get(21); internal_static_hbase_pb_DeleteTableResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_TruncateTableRequest_descriptor = - getDescriptor().getMessageTypes().get(24); + getDescriptor().getMessageTypes().get(22); internal_static_hbase_pb_TruncateTableRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_TruncateTableRequest_descriptor, new java.lang.String[] { "TableName", "PreserveSplits", "NonceGroup", "Nonce", }); internal_static_hbase_pb_TruncateTableResponse_descriptor = - getDescriptor().getMessageTypes().get(25); + getDescriptor().getMessageTypes().get(23); internal_static_hbase_pb_TruncateTableResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_TruncateTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_EnableTableRequest_descriptor = - getDescriptor().getMessageTypes().get(26); + getDescriptor().getMessageTypes().get(24); internal_static_hbase_pb_EnableTableRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableTableRequest_descriptor, new java.lang.String[] { "TableName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_EnableTableResponse_descriptor = - getDescriptor().getMessageTypes().get(27); + getDescriptor().getMessageTypes().get(25); internal_static_hbase_pb_EnableTableResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DisableTableRequest_descriptor = - getDescriptor().getMessageTypes().get(28); + getDescriptor().getMessageTypes().get(26); internal_static_hbase_pb_DisableTableRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DisableTableRequest_descriptor, new java.lang.String[] { "TableName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DisableTableResponse_descriptor = - getDescriptor().getMessageTypes().get(29); + getDescriptor().getMessageTypes().get(27); internal_static_hbase_pb_DisableTableResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DisableTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_ModifyTableRequest_descriptor = - getDescriptor().getMessageTypes().get(30); + getDescriptor().getMessageTypes().get(28); internal_static_hbase_pb_ModifyTableRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyTableRequest_descriptor, new java.lang.String[] { "TableName", "TableSchema", "NonceGroup", "Nonce", }); internal_static_hbase_pb_ModifyTableResponse_descriptor = - getDescriptor().getMessageTypes().get(31); + getDescriptor().getMessageTypes().get(29); internal_static_hbase_pb_ModifyTableResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_CreateNamespaceRequest_descriptor = - getDescriptor().getMessageTypes().get(32); + getDescriptor().getMessageTypes().get(30); internal_static_hbase_pb_CreateNamespaceRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceDescriptor", "NonceGroup", "Nonce", }); internal_static_hbase_pb_CreateNamespaceResponse_descriptor = - getDescriptor().getMessageTypes().get(33); + getDescriptor().getMessageTypes().get(31); internal_static_hbase_pb_CreateNamespaceResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateNamespaceResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DeleteNamespaceRequest_descriptor = - getDescriptor().getMessageTypes().get(34); + getDescriptor().getMessageTypes().get(32); internal_static_hbase_pb_DeleteNamespaceRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DeleteNamespaceResponse_descriptor = - getDescriptor().getMessageTypes().get(35); + getDescriptor().getMessageTypes().get(33); internal_static_hbase_pb_DeleteNamespaceResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteNamespaceResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_ModifyNamespaceRequest_descriptor = - getDescriptor().getMessageTypes().get(36); + getDescriptor().getMessageTypes().get(34); internal_static_hbase_pb_ModifyNamespaceRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceDescriptor", "NonceGroup", "Nonce", }); internal_static_hbase_pb_ModifyNamespaceResponse_descriptor = - getDescriptor().getMessageTypes().get(37); + getDescriptor().getMessageTypes().get(35); internal_static_hbase_pb_ModifyNamespaceResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyNamespaceResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor = - getDescriptor().getMessageTypes().get(38); + getDescriptor().getMessageTypes().get(36); internal_static_hbase_pb_GetNamespaceDescriptorRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor = - getDescriptor().getMessageTypes().get(39); + getDescriptor().getMessageTypes().get(37); internal_static_hbase_pb_GetNamespaceDescriptorResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor, new java.lang.String[] { "NamespaceDescriptor", }); internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor = - getDescriptor().getMessageTypes().get(40); + getDescriptor().getMessageTypes().get(38); internal_static_hbase_pb_ListNamespaceDescriptorsRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor = - getDescriptor().getMessageTypes().get(41); + getDescriptor().getMessageTypes().get(39); internal_static_hbase_pb_ListNamespaceDescriptorsResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor, new java.lang.String[] { "NamespaceDescriptor", }); internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor = - getDescriptor().getMessageTypes().get(42); + getDescriptor().getMessageTypes().get(40); internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor = - getDescriptor().getMessageTypes().get(43); + getDescriptor().getMessageTypes().get(41); internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor, new java.lang.String[] { "TableSchema", }); internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor = - getDescriptor().getMessageTypes().get(44); + getDescriptor().getMessageTypes().get(42); internal_static_hbase_pb_ListTableNamesByNamespaceRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor = - getDescriptor().getMessageTypes().get(45); + getDescriptor().getMessageTypes().get(43); internal_static_hbase_pb_ListTableNamesByNamespaceResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_ShutdownRequest_descriptor = - getDescriptor().getMessageTypes().get(46); + getDescriptor().getMessageTypes().get(44); internal_static_hbase_pb_ShutdownRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ShutdownRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ShutdownResponse_descriptor = - getDescriptor().getMessageTypes().get(47); + getDescriptor().getMessageTypes().get(45); internal_static_hbase_pb_ShutdownResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ShutdownResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_StopMasterRequest_descriptor = - getDescriptor().getMessageTypes().get(48); + getDescriptor().getMessageTypes().get(46); internal_static_hbase_pb_StopMasterRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_StopMasterRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_StopMasterResponse_descriptor = - getDescriptor().getMessageTypes().get(49); + getDescriptor().getMessageTypes().get(47); internal_static_hbase_pb_StopMasterResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_StopMasterResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor = - getDescriptor().getMessageTypes().get(50); + getDescriptor().getMessageTypes().get(48); internal_static_hbase_pb_IsInMaintenanceModeRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor = - getDescriptor().getMessageTypes().get(51); + getDescriptor().getMessageTypes().get(49); internal_static_hbase_pb_IsInMaintenanceModeResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor, new java.lang.String[] { "InMaintenanceMode", }); internal_static_hbase_pb_BalanceRequest_descriptor = - getDescriptor().getMessageTypes().get(52); + getDescriptor().getMessageTypes().get(50); internal_static_hbase_pb_BalanceRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalanceRequest_descriptor, new java.lang.String[] { "Force", }); internal_static_hbase_pb_BalanceResponse_descriptor = - getDescriptor().getMessageTypes().get(53); + getDescriptor().getMessageTypes().get(51); internal_static_hbase_pb_BalanceResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalanceResponse_descriptor, new java.lang.String[] { "BalancerRan", }); internal_static_hbase_pb_SetBalancerRunningRequest_descriptor = - getDescriptor().getMessageTypes().get(54); + getDescriptor().getMessageTypes().get(52); internal_static_hbase_pb_SetBalancerRunningRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetBalancerRunningRequest_descriptor, new java.lang.String[] { "On", "Synchronous", }); internal_static_hbase_pb_SetBalancerRunningResponse_descriptor = - getDescriptor().getMessageTypes().get(55); + getDescriptor().getMessageTypes().get(53); internal_static_hbase_pb_SetBalancerRunningResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetBalancerRunningResponse_descriptor, new java.lang.String[] { "PrevBalanceValue", }); internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor = - getDescriptor().getMessageTypes().get(56); + getDescriptor().getMessageTypes().get(54); internal_static_hbase_pb_IsBalancerEnabledRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor = - getDescriptor().getMessageTypes().get(57); + getDescriptor().getMessageTypes().get(55); internal_static_hbase_pb_IsBalancerEnabledResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor = - getDescriptor().getMessageTypes().get(58); + getDescriptor().getMessageTypes().get(56); internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor, new java.lang.String[] { "Enabled", "Synchronous", "SwitchTypes", }); internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor = - getDescriptor().getMessageTypes().get(59); + getDescriptor().getMessageTypes().get(57); internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor, new java.lang.String[] { "PrevValue", }); internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor = - getDescriptor().getMessageTypes().get(60); + getDescriptor().getMessageTypes().get(58); internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor, new java.lang.String[] { "SwitchType", }); internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor = - getDescriptor().getMessageTypes().get(61); + getDescriptor().getMessageTypes().get(59); internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_NormalizeRequest_descriptor = - getDescriptor().getMessageTypes().get(62); + getDescriptor().getMessageTypes().get(60); internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_NormalizeRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_NormalizeResponse_descriptor = - getDescriptor().getMessageTypes().get(63); + getDescriptor().getMessageTypes().get(61); internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_NormalizeResponse_descriptor, new java.lang.String[] { "NormalizerRan", }); internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor = - getDescriptor().getMessageTypes().get(64); + getDescriptor().getMessageTypes().get(62); internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor, new java.lang.String[] { "On", }); internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor = - getDescriptor().getMessageTypes().get(65); + getDescriptor().getMessageTypes().get(63); internal_static_hbase_pb_SetNormalizerRunningResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor, new java.lang.String[] { "PrevNormalizerValue", }); internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor = - getDescriptor().getMessageTypes().get(66); + getDescriptor().getMessageTypes().get(64); internal_static_hbase_pb_IsNormalizerEnabledRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor = - getDescriptor().getMessageTypes().get(67); + getDescriptor().getMessageTypes().get(65); internal_static_hbase_pb_IsNormalizerEnabledResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_RunCatalogScanRequest_descriptor = - getDescriptor().getMessageTypes().get(68); + getDescriptor().getMessageTypes().get(66); internal_static_hbase_pb_RunCatalogScanRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCatalogScanRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RunCatalogScanResponse_descriptor = - getDescriptor().getMessageTypes().get(69); + getDescriptor().getMessageTypes().get(67); internal_static_hbase_pb_RunCatalogScanResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCatalogScanResponse_descriptor, new java.lang.String[] { "ScanResult", }); internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor = - getDescriptor().getMessageTypes().get(70); + getDescriptor().getMessageTypes().get(68); internal_static_hbase_pb_EnableCatalogJanitorRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor, new java.lang.String[] { "Enable", }); internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor = - getDescriptor().getMessageTypes().get(71); + getDescriptor().getMessageTypes().get(69); internal_static_hbase_pb_EnableCatalogJanitorResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor, new java.lang.String[] { "PrevValue", }); internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor = - getDescriptor().getMessageTypes().get(72); + getDescriptor().getMessageTypes().get(70); internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor = - getDescriptor().getMessageTypes().get(73); + getDescriptor().getMessageTypes().get(71); internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor, new java.lang.String[] { "Value", }); internal_static_hbase_pb_RunCleanerChoreRequest_descriptor = - getDescriptor().getMessageTypes().get(74); + getDescriptor().getMessageTypes().get(72); internal_static_hbase_pb_RunCleanerChoreRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCleanerChoreRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RunCleanerChoreResponse_descriptor = - getDescriptor().getMessageTypes().get(75); + getDescriptor().getMessageTypes().get(73); internal_static_hbase_pb_RunCleanerChoreResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCleanerChoreResponse_descriptor, new java.lang.String[] { "CleanerChoreRan", }); internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor = - getDescriptor().getMessageTypes().get(76); + getDescriptor().getMessageTypes().get(74); internal_static_hbase_pb_SetCleanerChoreRunningRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor, new java.lang.String[] { "On", }); internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor = - getDescriptor().getMessageTypes().get(77); + getDescriptor().getMessageTypes().get(75); internal_static_hbase_pb_SetCleanerChoreRunningResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor, new java.lang.String[] { "PrevValue", }); internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor = - getDescriptor().getMessageTypes().get(78); + getDescriptor().getMessageTypes().get(76); internal_static_hbase_pb_IsCleanerChoreEnabledRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor = - getDescriptor().getMessageTypes().get(79); + getDescriptor().getMessageTypes().get(77); internal_static_hbase_pb_IsCleanerChoreEnabledResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor, new java.lang.String[] { "Value", }); internal_static_hbase_pb_SnapshotRequest_descriptor = - getDescriptor().getMessageTypes().get(80); + getDescriptor().getMessageTypes().get(78); internal_static_hbase_pb_SnapshotRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SnapshotRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_SnapshotResponse_descriptor = - getDescriptor().getMessageTypes().get(81); + getDescriptor().getMessageTypes().get(79); internal_static_hbase_pb_SnapshotResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SnapshotResponse_descriptor, new java.lang.String[] { "ExpectedTimeout", }); internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor = - getDescriptor().getMessageTypes().get(82); + getDescriptor().getMessageTypes().get(80); internal_static_hbase_pb_GetCompletedSnapshotsRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor = - getDescriptor().getMessageTypes().get(83); + getDescriptor().getMessageTypes().get(81); internal_static_hbase_pb_GetCompletedSnapshotsResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor, new java.lang.String[] { "Snapshots", }); internal_static_hbase_pb_DeleteSnapshotRequest_descriptor = - getDescriptor().getMessageTypes().get(84); + getDescriptor().getMessageTypes().get(82); internal_static_hbase_pb_DeleteSnapshotRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteSnapshotRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_DeleteSnapshotResponse_descriptor = - getDescriptor().getMessageTypes().get(85); + getDescriptor().getMessageTypes().get(83); internal_static_hbase_pb_DeleteSnapshotResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteSnapshotResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RestoreSnapshotRequest_descriptor = - getDescriptor().getMessageTypes().get(86); + getDescriptor().getMessageTypes().get(84); internal_static_hbase_pb_RestoreSnapshotRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RestoreSnapshotRequest_descriptor, new java.lang.String[] { "Snapshot", "NonceGroup", "Nonce", "RestoreACL", }); internal_static_hbase_pb_RestoreSnapshotResponse_descriptor = - getDescriptor().getMessageTypes().get(87); + getDescriptor().getMessageTypes().get(85); internal_static_hbase_pb_RestoreSnapshotResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RestoreSnapshotResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor = - getDescriptor().getMessageTypes().get(88); + getDescriptor().getMessageTypes().get(86); internal_static_hbase_pb_IsSnapshotDoneRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor = - getDescriptor().getMessageTypes().get(89); + getDescriptor().getMessageTypes().get(87); internal_static_hbase_pb_IsSnapshotDoneResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor, new java.lang.String[] { "Done", "Snapshot", }); internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor = - getDescriptor().getMessageTypes().get(90); + getDescriptor().getMessageTypes().get(88); internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor = - getDescriptor().getMessageTypes().get(91); + getDescriptor().getMessageTypes().get(89); internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor, new java.lang.String[] { "Done", }); internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor = - getDescriptor().getMessageTypes().get(92); + getDescriptor().getMessageTypes().get(90); internal_static_hbase_pb_GetSchemaAlterStatusRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor = - getDescriptor().getMessageTypes().get(93); + getDescriptor().getMessageTypes().get(91); internal_static_hbase_pb_GetSchemaAlterStatusResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor, new java.lang.String[] { "YetToUpdateRegions", "TotalRegions", }); internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor = - getDescriptor().getMessageTypes().get(94); + getDescriptor().getMessageTypes().get(92); internal_static_hbase_pb_GetTableDescriptorsRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor, new java.lang.String[] { "TableNames", "Regex", "IncludeSysTables", "Namespace", }); internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor = - getDescriptor().getMessageTypes().get(95); + getDescriptor().getMessageTypes().get(93); internal_static_hbase_pb_GetTableDescriptorsResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor, new java.lang.String[] { "TableSchema", }); internal_static_hbase_pb_GetTableNamesRequest_descriptor = - getDescriptor().getMessageTypes().get(96); + getDescriptor().getMessageTypes().get(94); internal_static_hbase_pb_GetTableNamesRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableNamesRequest_descriptor, new java.lang.String[] { "Regex", "IncludeSysTables", "Namespace", }); internal_static_hbase_pb_GetTableNamesResponse_descriptor = - getDescriptor().getMessageTypes().get(97); + getDescriptor().getMessageTypes().get(95); internal_static_hbase_pb_GetTableNamesResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableNamesResponse_descriptor, new java.lang.String[] { "TableNames", }); internal_static_hbase_pb_GetTableStateRequest_descriptor = - getDescriptor().getMessageTypes().get(98); + getDescriptor().getMessageTypes().get(96); internal_static_hbase_pb_GetTableStateRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableStateRequest_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_GetTableStateResponse_descriptor = - getDescriptor().getMessageTypes().get(99); + getDescriptor().getMessageTypes().get(97); internal_static_hbase_pb_GetTableStateResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableStateResponse_descriptor, new java.lang.String[] { "TableState", }); internal_static_hbase_pb_GetClusterStatusRequest_descriptor = - getDescriptor().getMessageTypes().get(100); + getDescriptor().getMessageTypes().get(98); internal_static_hbase_pb_GetClusterStatusRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetClusterStatusRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetClusterStatusResponse_descriptor = - getDescriptor().getMessageTypes().get(101); + getDescriptor().getMessageTypes().get(99); internal_static_hbase_pb_GetClusterStatusResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetClusterStatusResponse_descriptor, new java.lang.String[] { "ClusterStatus", }); internal_static_hbase_pb_IsMasterRunningRequest_descriptor = - getDescriptor().getMessageTypes().get(102); + getDescriptor().getMessageTypes().get(100); internal_static_hbase_pb_IsMasterRunningRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsMasterRunningRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsMasterRunningResponse_descriptor = - getDescriptor().getMessageTypes().get(103); + getDescriptor().getMessageTypes().get(101); internal_static_hbase_pb_IsMasterRunningResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsMasterRunningResponse_descriptor, new java.lang.String[] { "IsMasterRunning", }); internal_static_hbase_pb_ExecProcedureRequest_descriptor = - getDescriptor().getMessageTypes().get(104); + getDescriptor().getMessageTypes().get(102); internal_static_hbase_pb_ExecProcedureRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ExecProcedureRequest_descriptor, new java.lang.String[] { "Procedure", }); internal_static_hbase_pb_ExecProcedureResponse_descriptor = - getDescriptor().getMessageTypes().get(105); + getDescriptor().getMessageTypes().get(103); internal_static_hbase_pb_ExecProcedureResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ExecProcedureResponse_descriptor, new java.lang.String[] { "ExpectedTimeout", "ReturnData", }); internal_static_hbase_pb_IsProcedureDoneRequest_descriptor = - getDescriptor().getMessageTypes().get(106); + getDescriptor().getMessageTypes().get(104); internal_static_hbase_pb_IsProcedureDoneRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsProcedureDoneRequest_descriptor, new java.lang.String[] { "Procedure", }); internal_static_hbase_pb_IsProcedureDoneResponse_descriptor = - getDescriptor().getMessageTypes().get(107); + getDescriptor().getMessageTypes().get(105); internal_static_hbase_pb_IsProcedureDoneResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsProcedureDoneResponse_descriptor, new java.lang.String[] { "Done", "Snapshot", }); internal_static_hbase_pb_GetProcedureResultRequest_descriptor = - getDescriptor().getMessageTypes().get(108); + getDescriptor().getMessageTypes().get(106); internal_static_hbase_pb_GetProcedureResultRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetProcedureResultRequest_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_GetProcedureResultResponse_descriptor = - getDescriptor().getMessageTypes().get(109); + getDescriptor().getMessageTypes().get(107); internal_static_hbase_pb_GetProcedureResultResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetProcedureResultResponse_descriptor, new java.lang.String[] { "State", "SubmittedTime", "LastUpdate", "Result", "Exception", }); internal_static_hbase_pb_AbortProcedureRequest_descriptor = - getDescriptor().getMessageTypes().get(110); + getDescriptor().getMessageTypes().get(108); internal_static_hbase_pb_AbortProcedureRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AbortProcedureRequest_descriptor, new java.lang.String[] { "ProcId", "MayInterruptIfRunning", }); internal_static_hbase_pb_AbortProcedureResponse_descriptor = - getDescriptor().getMessageTypes().get(111); + getDescriptor().getMessageTypes().get(109); internal_static_hbase_pb_AbortProcedureResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AbortProcedureResponse_descriptor, new java.lang.String[] { "IsProcedureAborted", }); internal_static_hbase_pb_ListProceduresRequest_descriptor = - getDescriptor().getMessageTypes().get(112); + getDescriptor().getMessageTypes().get(110); internal_static_hbase_pb_ListProceduresRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListProceduresRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListProceduresResponse_descriptor = - getDescriptor().getMessageTypes().get(113); + getDescriptor().getMessageTypes().get(111); internal_static_hbase_pb_ListProceduresResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListProceduresResponse_descriptor, new java.lang.String[] { "Procedure", }); internal_static_hbase_pb_ListLocksRequest_descriptor = - getDescriptor().getMessageTypes().get(114); + getDescriptor().getMessageTypes().get(112); internal_static_hbase_pb_ListLocksRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListLocksRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListLocksResponse_descriptor = - getDescriptor().getMessageTypes().get(115); + getDescriptor().getMessageTypes().get(113); internal_static_hbase_pb_ListLocksResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListLocksResponse_descriptor, new java.lang.String[] { "Lock", }); internal_static_hbase_pb_SetQuotaRequest_descriptor = - getDescriptor().getMessageTypes().get(116); + getDescriptor().getMessageTypes().get(114); internal_static_hbase_pb_SetQuotaRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetQuotaRequest_descriptor, new java.lang.String[] { "UserName", "UserGroup", "Namespace", "TableName", "RemoveAll", "BypassGlobals", "Throttle", "SpaceLimit", }); internal_static_hbase_pb_SetQuotaResponse_descriptor = - getDescriptor().getMessageTypes().get(117); + getDescriptor().getMessageTypes().get(115); internal_static_hbase_pb_SetQuotaResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetQuotaResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor = - getDescriptor().getMessageTypes().get(118); + getDescriptor().getMessageTypes().get(116); internal_static_hbase_pb_MajorCompactionTimestampRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor = - getDescriptor().getMessageTypes().get(119); + getDescriptor().getMessageTypes().get(117); internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor, new java.lang.String[] { "Region", }); internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor = - getDescriptor().getMessageTypes().get(120); + getDescriptor().getMessageTypes().get(118); internal_static_hbase_pb_MajorCompactionTimestampResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor, new java.lang.String[] { "CompactionTimestamp", }); internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor = - getDescriptor().getMessageTypes().get(121); + getDescriptor().getMessageTypes().get(119); internal_static_hbase_pb_SecurityCapabilitiesRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor = - getDescriptor().getMessageTypes().get(122); + getDescriptor().getMessageTypes().get(120); internal_static_hbase_pb_SecurityCapabilitiesResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor, new java.lang.String[] { "Capabilities", }); internal_static_hbase_pb_ListDrainingRegionServersRequest_descriptor = - getDescriptor().getMessageTypes().get(123); + getDescriptor().getMessageTypes().get(121); internal_static_hbase_pb_ListDrainingRegionServersRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListDrainingRegionServersRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListDrainingRegionServersResponse_descriptor = - getDescriptor().getMessageTypes().get(124); + getDescriptor().getMessageTypes().get(122); internal_static_hbase_pb_ListDrainingRegionServersResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListDrainingRegionServersResponse_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_DrainRegionServersRequest_descriptor = - getDescriptor().getMessageTypes().get(125); + getDescriptor().getMessageTypes().get(123); internal_static_hbase_pb_DrainRegionServersRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DrainRegionServersRequest_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_DrainRegionServersResponse_descriptor = - getDescriptor().getMessageTypes().get(126); + getDescriptor().getMessageTypes().get(124); internal_static_hbase_pb_DrainRegionServersResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DrainRegionServersResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RemoveDrainFromRegionServersRequest_descriptor = - getDescriptor().getMessageTypes().get(127); + getDescriptor().getMessageTypes().get(125); internal_static_hbase_pb_RemoveDrainFromRegionServersRequest_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RemoveDrainFromRegionServersRequest_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_RemoveDrainFromRegionServersResponse_descriptor = - getDescriptor().getMessageTypes().get(128); + getDescriptor().getMessageTypes().get(126); internal_static_hbase_pb_RemoveDrainFromRegionServersResponse_fieldAccessorTable = new org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RemoveDrainFromRegionServersResponse_descriptor, diff --git a/hbase-protocol-shaded/src/main/protobuf/Admin.proto b/hbase-protocol-shaded/src/main/protobuf/Admin.proto index 6d67c89..3b8891b 100644 --- a/hbase-protocol-shaded/src/main/protobuf/Admin.proto +++ b/hbase-protocol-shaded/src/main/protobuf/Admin.proto @@ -273,22 +273,6 @@ message ExecuteProceduresResponse { repeated CloseRegionResponse close_region = 2; } -/** - * Merges the specified regions. - *

- * This method currently closes the regions and then merges them - */ -message MergeRegionsRequest { - required RegionSpecifier region_a = 1; - required RegionSpecifier region_b = 2; - optional bool forcible = 3 [default = false]; - // wall clock time from master - optional uint64 master_system_time = 4; -} - -message MergeRegionsResponse { -} - service AdminService { rpc GetRegionInfo(GetRegionInfoRequest) returns(GetRegionInfoResponse); @@ -350,7 +334,4 @@ service AdminService { rpc ExecuteProcedures(ExecuteProceduresRequest) returns(ExecuteProceduresResponse); - - rpc MergeRegions(MergeRegionsRequest) - returns(MergeRegionsResponse); } diff --git a/hbase-protocol-shaded/src/main/protobuf/Master.proto b/hbase-protocol-shaded/src/main/protobuf/Master.proto index c9c586f..8d7cad9 100644 --- a/hbase-protocol-shaded/src/main/protobuf/Master.proto +++ b/hbase-protocol-shaded/src/main/protobuf/Master.proto @@ -81,20 +81,6 @@ message MoveRegionRequest { message MoveRegionResponse { } - /** - * Dispatch merging the specified regions. - */ -message DispatchMergingRegionsRequest { - required RegionSpecifier region_a = 1; - required RegionSpecifier region_b = 2; - optional bool forcible = 3 [default = false]; - optional uint64 nonce_group = 4 [default = 0]; - optional uint64 nonce = 5 [default = 0]; -} - -message DispatchMergingRegionsResponse { - optional uint64 proc_id = 1; -} /** * Merging the specified regions in a table. @@ -667,10 +653,6 @@ service MasterService { rpc ModifyColumn(ModifyColumnRequest) returns(ModifyColumnResponse); - /** Master dispatch merging the regions */ - rpc DispatchMergingRegions(DispatchMergingRegionsRequest) - returns(DispatchMergingRegionsResponse); - /** Move the region region to the destination server. */ rpc MoveRegion(MoveRegionRequest) returns(MoveRegionResponse); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java index 67fda75..3b40b3c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java @@ -115,7 +115,6 @@ import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure; import org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure; import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure; import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure; -import org.apache.hadoop.hbase.master.procedure.DispatchMergingRegionsProcedure; import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure; import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants; import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; @@ -1531,59 +1530,6 @@ public class HMaster extends HRegionServer implements MasterServices { } @Override - public long dispatchMergingRegions( - final HRegionInfo regionInfoA, - final HRegionInfo regionInfoB, - final boolean forcible, - final long nonceGroup, - final long nonce) throws IOException { - checkInitialized(); - - TableName tableName = regionInfoA.getTable(); - if (tableName == null || regionInfoB.getTable() == null) { - throw new UnknownRegionException ("Can't merge regions without table associated"); - } - - if (!tableName.equals(regionInfoB.getTable())) { - throw new IOException ("Cannot merge regions from two different tables"); - } - - if (regionInfoA.compareTo(regionInfoB) == 0) { - throw new MergeRegionException( - "Cannot merge a region to itself " + regionInfoA + ", " + regionInfoB); - } - - final HRegionInfo [] regionsToMerge = new HRegionInfo[2]; - regionsToMerge [0] = regionInfoA; - regionsToMerge [1] = regionInfoB; - - return MasterProcedureUtil.submitProcedure( - new MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) { - @Override - protected void run() throws IOException { - MasterCoprocessorHost mcph = getMaster().getMasterCoprocessorHost(); - if (mcph != null) { - mcph.preDispatchMerge(regionInfoA, regionInfoB); - } - - LOG.info(getClientIdAuditPrefix() + " Dispatch merge regions " + - regionsToMerge[0].getEncodedName() + " and " + regionsToMerge[1].getEncodedName()); - - submitProcedure(new DispatchMergingRegionsProcedure( - procedureExecutor.getEnvironment(), tableName, regionsToMerge, forcible)); - if (mcph != null) { - mcph.postDispatchMerge(regionInfoA, regionInfoB); - } - } - - @Override - protected String getDescription() { - return "DispatchMergingRegionsProcedure"; - } - }); - } - - @Override public long mergeRegions( final HRegionInfo[] regionsToMerge, final boolean forcible, diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java index 4d18ac9..6064f9b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java @@ -810,28 +810,6 @@ public class MasterCoprocessorHost }); } - public void preDispatchMerge(final HRegionInfo regionInfoA, final HRegionInfo regionInfoB) - throws IOException { - execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { - @Override - public void call(MasterObserver oserver, ObserverContext ctx) - throws IOException { - oserver.preDispatchMerge(ctx, regionInfoA, regionInfoB); - } - }); - } - - public void postDispatchMerge(final HRegionInfo regionInfoA, final HRegionInfo regionInfoB) - throws IOException { - execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { - @Override - public void call(MasterObserver oserver, ObserverContext ctx) - throws IOException { - oserver.postDispatchMerge(ctx, regionInfoA, regionInfoB); - } - }); - } - public void preMergeRegions(final HRegionInfo[] regionsToMerge) throws IOException { execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java index 313c9c8..392e52f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java @@ -2009,34 +2009,4 @@ public class MasterRpcServices extends RSRpcServices throw new ServiceException(e); } } - - @Override - public DispatchMergingRegionsResponse dispatchMergingRegions(RpcController controller, - DispatchMergingRegionsRequest request) throws ServiceException { - final byte[] encodedNameOfRegionA = request.getRegionA().getValue().toByteArray(); - final byte[] encodedNameOfRegionB = request.getRegionB().getValue().toByteArray(); - if (request.getRegionA().getType() != RegionSpecifierType.ENCODED_REGION_NAME || - request.getRegionB().getType() != RegionSpecifierType.ENCODED_REGION_NAME) { - LOG.warn("mergeRegions specifier type: expected: " + RegionSpecifierType.ENCODED_REGION_NAME + - " actual: region_a=" + - request.getRegionA().getType() + ", region_b=" + - request.getRegionB().getType()); - } - RegionStates regionStates = master.getAssignmentManager().getRegionStates(); - RegionState regionStateA = regionStates.getRegionState(Bytes.toString(encodedNameOfRegionA)); - RegionState regionStateB = regionStates.getRegionState(Bytes.toString(encodedNameOfRegionB)); - if (regionStateA == null || regionStateB == null) { - throw new ServiceException(new UnknownRegionException( - Bytes.toStringBinary(regionStateA == null? encodedNameOfRegionA: encodedNameOfRegionB))); - } - final HRegionInfo regionInfoA = regionStateA.getRegion(); - final HRegionInfo regionInfoB = regionStateB.getRegion(); - try { - long procId = master.dispatchMergingRegions(regionInfoA, regionInfoB, request.getForcible(), - request.getNonceGroup(), request.getNonce()); - return DispatchMergingRegionsResponse.newBuilder().setProcId(procId).build(); - } catch (IOException ioe) { - throw new ServiceException(ioe); - } - } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java index c515435..7a4dcc0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java @@ -282,23 +282,6 @@ public interface MasterServices extends Server { throws IOException; /** - * Merge two regions. The real implementation is on the regionserver, master - * just move the regions together and send MERGE RPC to regionserver - * @param region_a region to merge - * @param region_b region to merge - * @param forcible true if do a compulsory merge, otherwise we will only merge - * two adjacent regions - * @return procedure Id - * @throws IOException - */ - long dispatchMergingRegions( - final HRegionInfo region_a, - final HRegionInfo region_b, - final boolean forcible, - final long nonceGroup, - final long nonce) throws IOException; - - /** * Merge regions in a table. * @param regionsToMerge daughter regions to merge * @param forcible whether to force to merge even two regions are not adjacent diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java index 4dff6f4..a187fbc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java @@ -622,37 +622,6 @@ public class ServerManager { } } - /** - * Sends an MERGE REGIONS RPC to the specified server to merge the specified - * regions. - *

- * A region server could reject the close request because it either does not - * have the specified region. - * @param server server to merge regions - * @param region_a region to merge - * @param region_b region to merge - * @param forcible true if do a compulsory merge, otherwise we will only merge - * two adjacent regions - * @throws IOException - */ - public void sendRegionsMerge(ServerName server, HRegionInfo region_a, - HRegionInfo region_b, boolean forcible, final User user) throws IOException { - if (server == null) - throw new NullPointerException("Passed server is null"); - if (region_a == null || region_b == null) - throw new NullPointerException("Passed region is null"); - AdminService.BlockingInterface admin = getRsAdmin(server); - if (admin == null) { - throw new IOException("Attempting to send MERGE REGIONS RPC to server " - + server.toString() + " for region " - + region_a.getRegionNameAsString() + "," - + region_b.getRegionNameAsString() - + " failed because no RPC connection found to this server"); - } - HBaseRpcController controller = newRpcController(); - ProtobufUtil.mergeRegions(controller, admin, region_a, region_b, forcible, user); - } - @VisibleForTesting public void moveFromOnlineToDeadServers(final ServerName sn) { synchronized (onlineServers) { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java deleted file mode 100644 index 15ed429..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.java +++ /dev/null @@ -1,584 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.master.procedure; - -import java.io.IOException; -import java.io.InputStream; -import java.io.InterruptedIOException; -import java.io.OutputStream; -import java.util.Map; -import java.util.concurrent.atomic.AtomicBoolean; - -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.hbase.HRegionInfo; -import org.apache.hadoop.hbase.RegionLoad; -import org.apache.hadoop.hbase.ServerLoad; -import org.apache.hadoop.hbase.ServerName; -import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.UnknownRegionException; -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.exceptions.MergeRegionException; -import org.apache.hadoop.hbase.exceptions.RegionOpeningException; -import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -import org.apache.hadoop.hbase.master.assignment.RegionStates; -import org.apache.hadoop.hbase.master.CatalogJanitor; -import org.apache.hadoop.hbase.master.RegionPlan; -import org.apache.hadoop.hbase.master.RegionState; -import org.apache.hadoop.hbase.master.ServerManager; -import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos; -import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.DispatchMergingRegionsState; -import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; - -/** - * The procedure to Merge a region in a table. - */ -@InterfaceAudience.Private -public class DispatchMergingRegionsProcedure - extends AbstractStateMachineTableProcedure { - private static final Log LOG = LogFactory.getLog(DispatchMergingRegionsProcedure.class); - - private final AtomicBoolean aborted = new AtomicBoolean(false); - private Boolean traceEnabled; - private AssignmentManager assignmentManager; - private int timeout; - private ServerName regionLocation; - private String regionsToMergeListFullName; - private String regionsToMergeListEncodedName; - - private TableName tableName; - private HRegionInfo [] regionsToMerge; - private boolean forcible; - - public DispatchMergingRegionsProcedure() { - this.traceEnabled = isTraceEnabled(); - this.assignmentManager = null; - this.timeout = -1; - this.regionLocation = null; - this.regionsToMergeListFullName = null; - this.regionsToMergeListEncodedName = null; - } - - public DispatchMergingRegionsProcedure( - final MasterProcedureEnv env, - final TableName tableName, - final HRegionInfo [] regionsToMerge, - final boolean forcible) { - super(env); - this.traceEnabled = isTraceEnabled(); - this.assignmentManager = getAssignmentManager(env); - this.tableName = tableName; - // For now, we only merge 2 regions. It could be extended to more than 2 regions in - // the future. - assert(regionsToMerge.length == 2); - this.regionsToMerge = regionsToMerge; - this.forcible = forcible; - - this.timeout = -1; - this.regionsToMergeListFullName = getRegionsToMergeListFullNameString(); - this.regionsToMergeListEncodedName = getRegionsToMergeListEncodedNameString(); - } - - @Override - protected Flow executeFromState( - final MasterProcedureEnv env, - final DispatchMergingRegionsState state) throws InterruptedException { - if (isTraceEnabled()) { - LOG.trace(this + " execute state=" + state); - } - - try { - switch (state) { - case DISPATCH_MERGING_REGIONS_PREPARE: - prepareMergeRegion(env); - setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_PRE_OPERATION); - break; - case DISPATCH_MERGING_REGIONS_PRE_OPERATION: - //Unused for now - reserve to add preMerge coprocessor in the future - setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS); - break; - case DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS: - if (MoveRegionsToSameRS(env)) { - setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS); - } else { - LOG.info("Cancel merging regions " + getRegionsToMergeListFullNameString() - + ", because can't move them to the same RS"); - setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_POST_OPERATION); - } - break; - case DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS: - doMergeInRS(env); - setNextState(DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_POST_OPERATION); - break; - case DISPATCH_MERGING_REGIONS_POST_OPERATION: - //Unused for now - reserve to add postCompletedMerge coprocessor in the future - return Flow.NO_MORE_STATE; - default: - throw new UnsupportedOperationException(this + " unhandled state=" + state); - } - } catch (IOException e) { - LOG.warn("Error trying to merge regions " + getRegionsToMergeListFullNameString() + - " in the table " + tableName + " (in state=" + state + ")", e); - - setFailure("master-merge-regions", e); - } - return Flow.HAS_MORE_STATE; - } - - @Override - protected void rollbackState( - final MasterProcedureEnv env, - final DispatchMergingRegionsState state) throws IOException, InterruptedException { - if (isTraceEnabled()) { - LOG.trace(this + " rollback state=" + state); - } - - try { - switch (state) { - case DISPATCH_MERGING_REGIONS_POST_OPERATION: - case DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS: - String msg = this + " We are in the " + state + " state." - + " It is complicated to rollback the merge operation that region server is working on." - + " Rollback is not supported and we should let the merge operation to complete"; - LOG.warn(msg); - // PONR - throw new UnsupportedOperationException(this + " unhandled state=" + state); - case DISPATCH_MERGING_REGIONS_MOVE_REGION_TO_SAME_RS: - break; // nothing to rollback - case DISPATCH_MERGING_REGIONS_PRE_OPERATION: - break; // nothing to rollback - case DISPATCH_MERGING_REGIONS_PREPARE: - break; // nothing to rollback - default: - throw new UnsupportedOperationException(this + " unhandled state=" + state); - } - } catch (Exception e) { - // This will be retried. Unless there is a bug in the code, - // this should be just a "temporary error" (e.g. network down) - LOG.warn("Failed rollback attempt step " + state + " for merging the regions " - + getRegionsToMergeListFullNameString() + " in table " + tableName, e); - throw e; - } - } - - @Override - protected DispatchMergingRegionsState getState(final int stateId) { - return DispatchMergingRegionsState.valueOf(stateId); - } - - @Override - protected int getStateId(final DispatchMergingRegionsState state) { - return state.getNumber(); - } - - @Override - protected DispatchMergingRegionsState getInitialState() { - return DispatchMergingRegionsState.DISPATCH_MERGING_REGIONS_PREPARE; - } - - /* - * Check whether we are in the state that can be rollback - */ - @Override - protected boolean isRollbackSupported(final DispatchMergingRegionsState state) { - switch (state) { - case DISPATCH_MERGING_REGIONS_POST_OPERATION: - case DISPATCH_MERGING_REGIONS_DO_MERGE_IN_RS: - // It is not safe to rollback if we reach to these states. - return false; - default: - break; - } - return true; - } - - @Override - public void serializeStateData(final OutputStream stream) throws IOException { - super.serializeStateData(stream); - - MasterProcedureProtos.DispatchMergingRegionsStateData.Builder dispatchMergingRegionsMsg = - MasterProcedureProtos.DispatchMergingRegionsStateData.newBuilder() - .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser())) - .setTableName(ProtobufUtil.toProtoTableName(tableName)) - .setForcible(forcible); - for (HRegionInfo hri: regionsToMerge) { - dispatchMergingRegionsMsg.addRegionInfo(HRegionInfo.convert(hri)); - } - dispatchMergingRegionsMsg.build().writeDelimitedTo(stream); - } - - @Override - public void deserializeStateData(final InputStream stream) throws IOException { - super.deserializeStateData(stream); - - MasterProcedureProtos.DispatchMergingRegionsStateData dispatchMergingRegionsMsg = - MasterProcedureProtos.DispatchMergingRegionsStateData.parseDelimitedFrom(stream); - setUser(MasterProcedureUtil.toUserInfo(dispatchMergingRegionsMsg.getUserInfo())); - tableName = ProtobufUtil.toTableName(dispatchMergingRegionsMsg.getTableName()); - - assert(dispatchMergingRegionsMsg.getRegionInfoCount() == 2); - regionsToMerge = new HRegionInfo[dispatchMergingRegionsMsg.getRegionInfoCount()]; - for (int i = 0; i < regionsToMerge.length; i++) { - regionsToMerge[i] = HRegionInfo.convert(dispatchMergingRegionsMsg.getRegionInfo(i)); - } - } - - @Override - public void toStringClassDetails(StringBuilder sb) { - sb.append(getClass().getSimpleName()); - sb.append(" (table="); - sb.append(tableName); - sb.append(" regions="); - sb.append(getRegionsToMergeListFullNameString()); - sb.append(" forcible="); - sb.append(forcible); - sb.append(")"); - } - - @Override - protected LockState acquireLock(final MasterProcedureEnv env) { - if (!getTableName().isSystemTable() && env.waitInitialized(this)) { - return LockState.LOCK_EVENT_WAIT; - } - if (env.getProcedureScheduler().waitRegions(this, getTableName(), regionsToMerge)) { - return LockState.LOCK_EVENT_WAIT; - } - return LockState.LOCK_ACQUIRED; - } - - @Override - protected void releaseLock(final MasterProcedureEnv env) { - env.getProcedureScheduler().wakeRegions(this, getTableName(), regionsToMerge[0], regionsToMerge[1]); - } - - @Override - public TableName getTableName() { - return tableName; - } - - @Override - public TableOperationType getTableOperationType() { - return TableOperationType.REGION_MERGE; - } - - /** - * Prepare merge and do some check - * @param env MasterProcedureEnv - * @throws IOException - */ - private void prepareMergeRegion(final MasterProcedureEnv env) throws IOException { - // Note: the following logic assumes that we only have 2 regions to merge. In the future, - // if we want to extend to more than 2 regions, the code needs to modify a little bit. - // - CatalogJanitor catalogJanitor = env.getMasterServices().getCatalogJanitor(); - boolean regionAHasMergeQualifier = !catalogJanitor.cleanMergeQualifier(regionsToMerge[0]); - if (regionAHasMergeQualifier - || !catalogJanitor.cleanMergeQualifier(regionsToMerge[1])) { - String msg = "Skip merging regions " + regionsToMerge[0].getRegionNameAsString() - + ", " + regionsToMerge[1].getRegionNameAsString() + ", because region " - + (regionAHasMergeQualifier ? regionsToMerge[0].getEncodedName() : regionsToMerge[1] - .getEncodedName()) + " has merge qualifier"; - LOG.info(msg); - throw new MergeRegionException(msg); - } - - RegionStates regionStates = getAssignmentManager(env).getRegionStates(); - RegionState regionStateA = regionStates.getRegionState(regionsToMerge[0].getEncodedName()); - RegionState regionStateB = regionStates.getRegionState(regionsToMerge[1].getEncodedName()); - if (regionStateA == null || regionStateB == null) { - throw new UnknownRegionException( - regionStateA == null ? - regionsToMerge[0].getEncodedName() : regionsToMerge[1].getEncodedName()); - } - - if (!regionStateA.isOpened() || !regionStateB.isOpened()) { - throw new MergeRegionException( - "Unable to merge regions not online " + regionStateA + ", " + regionStateB); - } - - if (regionsToMerge[0].getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID || - regionsToMerge[1].getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID) { - throw new MergeRegionException("Can't merge non-default replicas"); - } - - if (!forcible && !HRegionInfo.areAdjacent(regionsToMerge[0], regionsToMerge[1])) { - throw new MergeRegionException( - "Unable to merge not adjacent regions " - + regionsToMerge[0].getRegionNameAsString() + ", " - + regionsToMerge[1].getRegionNameAsString() - + " where forcible = " + forcible); - } - } - - /** - * Move all regions to the same region server - * @param env MasterProcedureEnv - * @return whether target regions hosted by the same RS - * @throws IOException - */ - private boolean MoveRegionsToSameRS(final MasterProcedureEnv env) throws IOException { - // Make sure regions are on the same regionserver before send merge - // regions request to region server. - // - boolean onSameRS = isRegionsOnTheSameServer(env); - if (!onSameRS) { - // Note: the following logic assumes that we only have 2 regions to merge. In the future, - // if we want to extend to more than 2 regions, the code needs to modify a little bit. - // - RegionStates regionStates = getAssignmentManager(env).getRegionStates(); - ServerName regionLocation2 = regionStates.getRegionServerOfRegion(regionsToMerge[1]); - - RegionLoad loadOfRegionA = getRegionLoad(env, regionLocation, regionsToMerge[0]); - RegionLoad loadOfRegionB = getRegionLoad(env, regionLocation2, regionsToMerge[1]); - if (loadOfRegionA != null && loadOfRegionB != null - && loadOfRegionA.getRequestsCount() < loadOfRegionB.getRequestsCount()) { - // switch regionsToMerge[0] and regionsToMerge[1] - HRegionInfo tmpRegion = this.regionsToMerge[0]; - this.regionsToMerge[0] = this.regionsToMerge[1]; - this.regionsToMerge[1] = tmpRegion; - ServerName tmpLocation = regionLocation; - regionLocation = regionLocation2; - regionLocation2 = tmpLocation; - } - - long startTime = EnvironmentEdgeManager.currentTime(); - - RegionPlan regionPlan = new RegionPlan(regionsToMerge[1], regionLocation2, regionLocation); - LOG.info("Moving regions to same server for merge: " + regionPlan.toString()); - getAssignmentManager(env).moveAsync(regionPlan); - do { - try { - Thread.sleep(20); - // Make sure check RIT first, then get region location, otherwise - // we would make a wrong result if region is online between getting - // region location and checking RIT - boolean isRIT = regionStates.isRegionInTransition(regionsToMerge[1]); - regionLocation2 = regionStates.getRegionServerOfRegion(regionsToMerge[1]); - onSameRS = regionLocation.equals(regionLocation2); - if (onSameRS || !isRIT) { - // Regions are on the same RS, or regionsToMerge[1] is not in - // RegionInTransition any more - break; - } - } catch (InterruptedException e) { - InterruptedIOException iioe = new InterruptedIOException(); - iioe.initCause(e); - throw iioe; - } - } while ((EnvironmentEdgeManager.currentTime() - startTime) <= getTimeout(env)); - } - return onSameRS; - } - - /** - * Do the real merge operation in the region server that hosts regions - * @param env MasterProcedureEnv - * @throws IOException - */ - private void doMergeInRS(final MasterProcedureEnv env) throws IOException { - long duration = 0; - long startTime = EnvironmentEdgeManager.currentTime(); - do { - try { - if (getServerName(env) == null) { - // The merge probably already happen. Check - RegionState regionState = getAssignmentManager(env).getRegionStates().getRegionState( - regionsToMerge[0].getEncodedName()); - if (regionState.isMerging() || regionState.isMerged()) { - LOG.info("Merge regions " + getRegionsToMergeListEncodedNameString() + - " is in progress or completed. No need to send a new request."); - } else { - LOG.warn("Cannot sending merge to hosting server of the regions " + - getRegionsToMergeListEncodedNameString() + " as the server is unknown"); - } - return; - } - // TODO: the following RPC call is not idempotent. Multiple calls (eg. after master - // failover, re-execute this step) could result in some exception thrown that does not - // paint the correct picture. This behavior is on-par with old releases. Improvement - // could happen in the future. - env.getMasterServices().getServerManager().sendRegionsMerge( - getServerName(env), - regionsToMerge[0], - regionsToMerge[1], - forcible, - getUser()); - LOG.info("Sent merge to server " + getServerName(env) + " for region " + - getRegionsToMergeListEncodedNameString() + ", forcible=" + forcible); - return; - } catch (RegionOpeningException roe) { - // Do a retry since region should be online on RS immediately - LOG.warn("Failed mergering regions in " + getServerName(env) + ", retrying...", roe); - } catch (Exception ie) { - LOG.warn("Failed sending merge to " + getServerName(env) + " for regions " + - getRegionsToMergeListEncodedNameString() + ", forcible=" + forcible, ie); - return; - } - } while ((duration = EnvironmentEdgeManager.currentTime() - startTime) <= getTimeout(env)); - - // If we reaches here, it means that we get timed out. - String msg = "Failed sending merge to " + getServerName(env) + " after " + duration + "ms"; - LOG.warn(msg); - throw new IOException(msg); - } - - private RegionLoad getRegionLoad( - final MasterProcedureEnv env, - final ServerName sn, - final HRegionInfo hri) { - ServerManager serverManager = env.getMasterServices().getServerManager(); - ServerLoad load = serverManager.getLoad(sn); - if (load != null) { - Map regionsLoad = load.getRegionsLoad(); - if (regionsLoad != null) { - return regionsLoad.get(hri.getRegionName()); - } - } - return null; - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @param env MasterProcedureEnv - * @return whether target regions hosted by the same RS - */ - private boolean isRegionsOnTheSameServer(final MasterProcedureEnv env) throws IOException{ - Boolean onSameRS = true; - int i = 0; - RegionStates regionStates = getAssignmentManager(env).getRegionStates(); - regionLocation = regionStates.getRegionServerOfRegion(regionsToMerge[i]); - if (regionLocation != null) { - for(i = 1; i < regionsToMerge.length; i++) { - ServerName regionLocation2 = regionStates.getRegionServerOfRegion(regionsToMerge[i]); - if (regionLocation2 != null) { - if (onSameRS) { - onSameRS = regionLocation.equals(regionLocation2); - } - } else { - // At least one region is not online, merge will fail, no need to continue. - break; - } - } - if (i == regionsToMerge.length) { - // Finish checking all regions, return the result; - return onSameRS; - } - } - - // If reaching here, at least one region is not online. - String msg = "Skip merging regions " + getRegionsToMergeListFullNameString() + - ", because region " + regionsToMerge[i].getEncodedName() + " is not online now."; - LOG.warn(msg); - throw new IOException(msg); - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @param env MasterProcedureEnv - * @return assignmentManager - */ - private AssignmentManager getAssignmentManager(final MasterProcedureEnv env) { - if (assignmentManager == null) { - assignmentManager = env.getMasterServices().getAssignmentManager(); - } - return assignmentManager; - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @param env MasterProcedureEnv - * @return timeout value - */ - private int getTimeout(final MasterProcedureEnv env) { - if (timeout == -1) { - timeout = env.getMasterConfiguration().getInt( - "hbase.master.regionmerge.timeout", regionsToMerge.length * 60 * 1000); - } - return timeout; - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @param env MasterProcedureEnv - * @return serverName - */ - private ServerName getServerName(final MasterProcedureEnv env) { - if (regionLocation == null) { - regionLocation = - getAssignmentManager(env).getRegionStates().getRegionServerOfRegion(regionsToMerge[0]); - } - return regionLocation; - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @param fullName whether return only encoded name - * @return region names in a list - */ - private String getRegionsToMergeListFullNameString() { - if (regionsToMergeListFullName == null) { - StringBuilder sb = new StringBuilder("["); - int i = 0; - while(i < regionsToMerge.length - 1) { - sb.append(regionsToMerge[i].getRegionNameAsString() + ", "); - i++; - } - sb.append(regionsToMerge[i].getRegionNameAsString() + " ]"); - regionsToMergeListFullName = sb.toString(); - } - return regionsToMergeListFullName; - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @return encoded region names - */ - private String getRegionsToMergeListEncodedNameString() { - if (regionsToMergeListEncodedName == null) { - StringBuilder sb = new StringBuilder("["); - int i = 0; - while(i < regionsToMerge.length - 1) { - sb.append(regionsToMerge[i].getEncodedName() + ", "); - i++; - } - sb.append(regionsToMerge[i].getEncodedName() + " ]"); - regionsToMergeListEncodedName = sb.toString(); - } - return regionsToMergeListEncodedName; - } - - /** - * The procedure could be restarted from a different machine. If the variable is null, we need to - * retrieve it. - * @return traceEnabled - */ - private Boolean isTraceEnabled() { - if (traceEnabled == null) { - traceEnabled = LOG.isTraceEnabled(); - } - return traceEnabled; - } -} diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java index 5d3b50b..9cc8087 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplit.java @@ -208,20 +208,6 @@ public class CompactSplit implements CompactionRequestor, PropagatingConfigurati return queueLists.toString(); } - public synchronized void requestRegionsMerge(final Region a, - final Region b, final boolean forcible, long masterSystemTime, User user) { - try { - mergePool.execute(new RegionMergeRequest(a, b, this.server, forcible, masterSystemTime,user)); - if (LOG.isDebugEnabled()) { - LOG.debug("Region merge requested for " + a + "," + b + ", forcible=" - + forcible + ". " + this); - } - } catch (RejectedExecutionException ree) { - LOG.warn("Could not execute merge for " + a + "," + b + ", forcible=" - + forcible, ree); - } - } - public synchronized boolean requestSplit(final Region r) { // don't split regions that are blocking if (shouldSplitRegion() && ((HRegion)r).getCompactPriority() >= Store.PRIORITY_USER) { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java index 8bd1b5d..d284a01 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java @@ -84,7 +84,6 @@ import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.VersionInfoUtil; import org.apache.hadoop.hbase.conf.ConfigurationObserver; import org.apache.hadoop.hbase.exceptions.FailedSanityCheckException; -import org.apache.hadoop.hbase.exceptions.MergeRegionException; import org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException; import org.apache.hadoop.hbase.exceptions.ScannerResetException; import org.apache.hadoop.hbase.filter.ByteArrayComparable; @@ -150,8 +149,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerIn import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse; -import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest; -import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse; @@ -3489,45 +3486,4 @@ public class RSRpcServices implements HBaseRPCErrorHandler, } return builder.build(); } - - /** - * Merge regions on the region server. - * - * @param controller the RPC controller - * @param request the request - * @return merge regions response - * @throws ServiceException - */ - @Override - @QosPriority(priority = HConstants.ADMIN_QOS) - // UNUSED AS OF AMv2 PURGE! - public MergeRegionsResponse mergeRegions(final RpcController controller, - final MergeRegionsRequest request) throws ServiceException { - try { - checkOpen(); - requestCount.increment(); - Region regionA = getRegion(request.getRegionA()); - Region regionB = getRegion(request.getRegionB()); - boolean forcible = request.getForcible(); - long masterSystemTime = request.hasMasterSystemTime() ? request.getMasterSystemTime() : -1; - regionA.startRegionOperation(Operation.MERGE_REGION); - regionB.startRegionOperation(Operation.MERGE_REGION); - if (regionA.getRegionInfo().getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID || - regionB.getRegionInfo().getReplicaId() != HRegionInfo.DEFAULT_REPLICA_ID) { - throw new ServiceException(new MergeRegionException("Can't merge non-default replicas")); - } - LOG.info("Receiving merging request for " + regionA + ", " + regionB - + ",forcible=" + forcible); - regionA.flush(true); - regionB.flush(true); - regionServer.compactSplitThread.requestRegionsMerge(regionA, regionB, forcible, - masterSystemTime, RpcServer.getRequestUser()); - return MergeRegionsResponse.newBuilder().build(); - } catch (DroppedSnapshotException ex) { - regionServer.abort("Replay of WAL required. Forcing server shutdown", ex); - throw new ServiceException(ex); - } catch (IOException ie) { - throw new ServiceException(ie); - } - } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java deleted file mode 100644 index e95932b..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeRequest.java +++ /dev/null @@ -1,108 +0,0 @@ -/** - * Copyright The Apache Software Foundation - * - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with this - * work for additional information regarding copyright ownership. The ASF - * licenses this file to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT - * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the - * License for the specific language governing permissions and limitations - * under the License. - */ -package org.apache.hadoop.hbase.regionserver; - -import java.security.PrivilegedAction; - -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.hbase.HRegionInfo; -import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode; - -import com.google.common.base.Preconditions; - -/** - * Handles processing region merges. Put in a queue, owned by HRegionServer. - */ -// UNUSED: REMOVE!!! -@InterfaceAudience.Private -class RegionMergeRequest implements Runnable { - private static final Log LOG = LogFactory.getLog(RegionMergeRequest.class); - private final HRegionInfo region_a; - private final HRegionInfo region_b; - private final HRegionServer server; - private final boolean forcible; - private final User user; - - RegionMergeRequest(Region a, Region b, HRegionServer hrs, boolean forcible, - long masterSystemTime, User user) { - Preconditions.checkNotNull(hrs); - this.region_a = a.getRegionInfo(); - this.region_b = b.getRegionInfo(); - this.server = hrs; - this.forcible = forcible; - this.user = user; - } - - @Override - public String toString() { - return "MergeRequest,regions:" + region_a + ", " + region_b + ", forcible=" - + forcible; - } - - private void doMerge() { - boolean success = false; - //server.metricsRegionServer.incrMergeRequest(); - - if (user != null && user.getUGI() != null) { - user.getUGI().doAs (new PrivilegedAction() { - @Override - public Void run() { - requestRegionMerge(); - return null; - } - }); - } else { - requestRegionMerge(); - } - } - - private void requestRegionMerge() { - final TableName table = region_a.getTable(); - if (!table.equals(region_b.getTable())) { - LOG.error("Can't merge regions from two different tables: " + region_a + ", " + region_b); - return; - } - - // TODO: fake merged region for compat with the report protocol - final HRegionInfo merged = new HRegionInfo(table); - - // Send the split request to the master. the master will do the validation on the split-key. - // The parent region will be unassigned and the two new regions will be assigned. - // hri_a and hri_b objects may not reflect the regions that will be created, those objectes - // are created just to pass the information to the reportRegionStateTransition(). - if (!server.reportRegionStateTransition(TransitionCode.READY_TO_MERGE, merged, region_a, region_b)) { - LOG.error("Unable to ask master to merge: " + region_a + ", " + region_b); - } - } - - @Override - public void run() { - if (this.server.isStopping() || this.server.isStopped()) { - LOG.debug("Skipping merge because server is stopping=" - + this.server.isStopping() + " or stopped=" + this.server.isStopped()); - return; - } - - doMerge(); - } -} diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java index 8f878b8..2d30d7e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockNoopMasterServices.java @@ -452,12 +452,6 @@ public class MockNoopMasterServices implements MasterServices, Server { } @Override - public long dispatchMergingRegions(HRegionInfo region_a, HRegionInfo region_b, boolean forcible, long nonceGroup, - long nonce) throws IOException { - return 0; - } - - @Override public ProcedureEvent getInitializedEvent() { // TODO Auto-generated method stub return null; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java index 7ac7571..fe67d21 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java @@ -69,8 +69,6 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerIn import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetServerInfoResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileResponse; -import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsRequest; -import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.MergeRegionsResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ReplicateWALEntryRequest; @@ -733,11 +731,4 @@ ClientProtos.ClientService.BlockingInterface, RegionServerServices { throws ServiceException { return null; } - - @Override - public MergeRegionsResponse mergeRegions(RpcController controller, MergeRegionsRequest request) - throws ServiceException { - // TODO Auto-generated method stub - return null; - } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicator.java index 0e69f1c..7b05326 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicator.java @@ -403,13 +403,6 @@ public class TestReplicator extends TestReplicationBase { throws ServiceException { return null; } - - @Override - public MergeRegionsResponse mergeRegions(RpcController controller, - MergeRegionsRequest request) - throws ServiceException { - return null; - } } public class FailureInjectingReplicatorForTest extends ReplicatorForTest { -- 2.10.1