required .RegionSpecifier region = 1;
+ */
+ boolean hasRegion();
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
+
+ // required .CoprocessorServiceCall call = 2;
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ boolean hasCall();
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall getCall();
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCallOrBuilder getCallOrBuilder();
+ }
/**
+ * Protobuf type {@code BatchCoprocessorServiceCall}
+ */
+ public static final class BatchCoprocessorServiceCall extends
+ com.google.protobuf.GeneratedMessage
+ implements BatchCoprocessorServiceCallOrBuilder {
+ // Use BatchCoprocessorServiceCall.newBuilder() to construct.
+ private BatchCoprocessorServiceCall(com.google.protobuf.GeneratedMessage.Builder> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private BatchCoprocessorServiceCall(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final BatchCoprocessorServiceCall defaultInstance;
+ public static BatchCoprocessorServiceCall getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public BatchCoprocessorServiceCall getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private BatchCoprocessorServiceCall(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ subBuilder = region_.toBuilder();
+ }
+ region_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(region_);
+ region_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000001;
+ break;
+ }
+ case 18: {
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ subBuilder = call_.toBuilder();
+ }
+ call_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(call_);
+ call_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000002;
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_BatchCoprocessorServiceCall_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_BatchCoprocessorServiceCall_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder.class);
+ }
+
+ public static com.google.protobuf.Parserrequired .RegionSpecifier region = 1;
+ */
+ public boolean hasRegion() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
+ return region_;
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
+ return region_;
+ }
+
+ // required .CoprocessorServiceCall call = 2;
+ public static final int CALL_FIELD_NUMBER = 2;
+ private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall call_;
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public boolean hasCall() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall getCall() {
+ return call_;
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCallOrBuilder getCallOrBuilder() {
+ return call_;
+ }
+
+ private void initFields() {
+ region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+ call_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.getDefaultInstance();
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasRegion()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!hasCall()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!getRegion().isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!getCall().isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeMessage(1, region_);
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeMessage(2, call_);
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(1, region_);
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(2, call_);
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall) obj;
+
+ boolean result = true;
+ result = result && (hasRegion() == other.hasRegion());
+ if (hasRegion()) {
+ result = result && getRegion()
+ .equals(other.getRegion());
+ }
+ result = result && (hasCall() == other.hasCall());
+ if (hasCall()) {
+ result = result && getCall()
+ .equals(other.getCall());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasRegion()) {
+ hash = (37 * hash) + REGION_FIELD_NUMBER;
+ hash = (53 * hash) + getRegion().hashCode();
+ }
+ if (hasCall()) {
+ hash = (37 * hash) + CALL_FIELD_NUMBER;
+ hash = (53 * hash) + getCall().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code BatchCoprocessorServiceCall}
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builderrequired .RegionSpecifier region = 1;
+ */
+ public boolean hasRegion() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
+ if (regionBuilder_ == null) {
+ return region_;
+ } else {
+ return regionBuilder_.getMessage();
+ }
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public Builder setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+ if (regionBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ region_ = value;
+ onChanged();
+ } else {
+ regionBuilder_.setMessage(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public Builder setRegion(
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
+ if (regionBuilder_ == null) {
+ region_ = builderForValue.build();
+ onChanged();
+ } else {
+ regionBuilder_.setMessage(builderForValue.build());
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public Builder mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+ if (regionBuilder_ == null) {
+ if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ region_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
+ region_ =
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial();
+ } else {
+ region_ = value;
+ }
+ onChanged();
+ } else {
+ regionBuilder_.mergeFrom(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public Builder clearRegion() {
+ if (regionBuilder_ == null) {
+ region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+ onChanged();
+ } else {
+ regionBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000001);
+ return this;
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() {
+ bitField0_ |= 0x00000001;
+ onChanged();
+ return getRegionFieldBuilder().getBuilder();
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
+ if (regionBuilder_ != null) {
+ return regionBuilder_.getMessageOrBuilder();
+ } else {
+ return region_;
+ }
+ }
+ /**
+ * required .RegionSpecifier region = 1;
+ */
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>
+ getRegionFieldBuilder() {
+ if (regionBuilder_ == null) {
+ regionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
+ region_,
+ getParentForChildren(),
+ isClean());
+ region_ = null;
+ }
+ return regionBuilder_;
+ }
+
+ // required .CoprocessorServiceCall call = 2;
+ private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall call_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.getDefaultInstance();
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCallOrBuilder> callBuilder_;
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public boolean hasCall() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall getCall() {
+ if (callBuilder_ == null) {
+ return call_;
+ } else {
+ return callBuilder_.getMessage();
+ }
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public Builder setCall(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall value) {
+ if (callBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ call_ = value;
+ onChanged();
+ } else {
+ callBuilder_.setMessage(value);
+ }
+ bitField0_ |= 0x00000002;
+ return this;
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public Builder setCall(
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.Builder builderForValue) {
+ if (callBuilder_ == null) {
+ call_ = builderForValue.build();
+ onChanged();
+ } else {
+ callBuilder_.setMessage(builderForValue.build());
+ }
+ bitField0_ |= 0x00000002;
+ return this;
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public Builder mergeCall(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall value) {
+ if (callBuilder_ == null) {
+ if (((bitField0_ & 0x00000002) == 0x00000002) &&
+ call_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.getDefaultInstance()) {
+ call_ =
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.newBuilder(call_).mergeFrom(value).buildPartial();
+ } else {
+ call_ = value;
+ }
+ onChanged();
+ } else {
+ callBuilder_.mergeFrom(value);
+ }
+ bitField0_ |= 0x00000002;
+ return this;
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public Builder clearCall() {
+ if (callBuilder_ == null) {
+ call_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.getDefaultInstance();
+ onChanged();
+ } else {
+ callBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000002);
+ return this;
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.Builder getCallBuilder() {
+ bitField0_ |= 0x00000002;
+ onChanged();
+ return getCallFieldBuilder().getBuilder();
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCallOrBuilder getCallOrBuilder() {
+ if (callBuilder_ != null) {
+ return callBuilder_.getMessageOrBuilder();
+ } else {
+ return call_;
+ }
+ }
+ /**
+ * required .CoprocessorServiceCall call = 2;
+ */
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCallOrBuilder>
+ getCallFieldBuilder() {
+ if (callBuilder_ == null) {
+ callBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCallOrBuilder>(
+ call_,
+ getParentForChildren(),
+ isClean());
+ call_ = null;
+ }
+ return callBuilder_;
+ }
+
+ // @@protoc_insertion_point(builder_scope:BatchCoprocessorServiceCall)
+ }
+
+ static {
+ defaultInstance = new BatchCoprocessorServiceCall(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:BatchCoprocessorServiceCall)
+ }
+
+ public interface BatchCoprocessorServiceRequestOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required string server_callback_name = 1;
+ /**
+ * required string server_callback_name = 1;
+ */
+ boolean hasServerCallbackName();
+ /**
+ * required string server_callback_name = 1;
+ */
+ java.lang.String getServerCallbackName();
+ /**
+ * required string server_callback_name = 1;
+ */
+ com.google.protobuf.ByteString
+ getServerCallbackNameBytes();
+
+ // optional bytes server_callback_initData = 2;
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ boolean hasServerCallbackInitData();
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ com.google.protobuf.ByteString getServerCallbackInitData();
+
+ // repeated .BatchCoprocessorServiceCall call = 3;
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ java.util.Listrepeated .BatchCoprocessorServiceCall call = 3;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall getCall(int index);
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ int getCallCount();
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ java.util.List extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCallOrBuilder>
+ getCallOrBuilderList();
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCallOrBuilder getCallOrBuilder(
+ int index);
+ }
+ /**
+ * Protobuf type {@code BatchCoprocessorServiceRequest}
+ */
+ public static final class BatchCoprocessorServiceRequest extends
+ com.google.protobuf.GeneratedMessage
+ implements BatchCoprocessorServiceRequestOrBuilder {
+ // Use BatchCoprocessorServiceRequest.newBuilder() to construct.
+ private BatchCoprocessorServiceRequest(com.google.protobuf.GeneratedMessage.Builder> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private BatchCoprocessorServiceRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final BatchCoprocessorServiceRequest defaultInstance;
+ public static BatchCoprocessorServiceRequest getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public BatchCoprocessorServiceRequest getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private BatchCoprocessorServiceRequest(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ serverCallbackName_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ bitField0_ |= 0x00000002;
+ serverCallbackInitData_ = input.readBytes();
+ break;
+ }
+ case 26: {
+ if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+ call_ = new java.util.ArrayListrequired string server_callback_name = 1;
+ */
+ public boolean hasServerCallbackName() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public java.lang.String getServerCallbackName() {
+ java.lang.Object ref = serverCallbackName_;
+ if (ref instanceof java.lang.String) {
+ return (java.lang.String) ref;
+ } else {
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ serverCallbackName_ = s;
+ }
+ return s;
+ }
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public com.google.protobuf.ByteString
+ getServerCallbackNameBytes() {
+ java.lang.Object ref = serverCallbackName_;
+ if (ref instanceof java.lang.String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ serverCallbackName_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
+ }
+ }
+
+ // optional bytes server_callback_initData = 2;
+ public static final int SERVER_CALLBACK_INITDATA_FIELD_NUMBER = 2;
+ private com.google.protobuf.ByteString serverCallbackInitData_;
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ public boolean hasServerCallbackInitData() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ public com.google.protobuf.ByteString getServerCallbackInitData() {
+ return serverCallbackInitData_;
+ }
+
+ // repeated .BatchCoprocessorServiceCall call = 3;
+ public static final int CALL_FIELD_NUMBER = 3;
+ private java.util.Listrepeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public java.util.Listrepeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public java.util.List extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCallOrBuilder>
+ getCallOrBuilderList() {
+ return call_;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public int getCallCount() {
+ return call_.size();
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall getCall(int index) {
+ return call_.get(index);
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCallOrBuilder getCallOrBuilder(
+ int index) {
+ return call_.get(index);
+ }
+
+ private void initFields() {
+ serverCallbackName_ = "";
+ serverCallbackInitData_ = com.google.protobuf.ByteString.EMPTY;
+ call_ = java.util.Collections.emptyList();
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasServerCallbackName()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ for (int i = 0; i < getCallCount(); i++) {
+ if (!getCall(i).isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeBytes(1, getServerCallbackNameBytes());
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeBytes(2, serverCallbackInitData_);
+ }
+ for (int i = 0; i < call_.size(); i++) {
+ output.writeMessage(3, call_.get(i));
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(1, getServerCallbackNameBytes());
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(2, serverCallbackInitData_);
+ }
+ for (int i = 0; i < call_.size(); i++) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(3, call_.get(i));
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest) obj;
+
+ boolean result = true;
+ result = result && (hasServerCallbackName() == other.hasServerCallbackName());
+ if (hasServerCallbackName()) {
+ result = result && getServerCallbackName()
+ .equals(other.getServerCallbackName());
+ }
+ result = result && (hasServerCallbackInitData() == other.hasServerCallbackInitData());
+ if (hasServerCallbackInitData()) {
+ result = result && getServerCallbackInitData()
+ .equals(other.getServerCallbackInitData());
+ }
+ result = result && getCallList()
+ .equals(other.getCallList());
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasServerCallbackName()) {
+ hash = (37 * hash) + SERVER_CALLBACK_NAME_FIELD_NUMBER;
+ hash = (53 * hash) + getServerCallbackName().hashCode();
+ }
+ if (hasServerCallbackInitData()) {
+ hash = (37 * hash) + SERVER_CALLBACK_INITDATA_FIELD_NUMBER;
+ hash = (53 * hash) + getServerCallbackInitData().hashCode();
+ }
+ if (getCallCount() > 0) {
+ hash = (37 * hash) + CALL_FIELD_NUMBER;
+ hash = (53 * hash) + getCallList().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code BatchCoprocessorServiceRequest}
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builderrequired string server_callback_name = 1;
+ */
+ public boolean hasServerCallbackName() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public java.lang.String getServerCallbackName() {
+ java.lang.Object ref = serverCallbackName_;
+ if (!(ref instanceof java.lang.String)) {
+ java.lang.String s = ((com.google.protobuf.ByteString) ref)
+ .toStringUtf8();
+ serverCallbackName_ = s;
+ return s;
+ } else {
+ return (java.lang.String) ref;
+ }
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public com.google.protobuf.ByteString
+ getServerCallbackNameBytes() {
+ java.lang.Object ref = serverCallbackName_;
+ if (ref instanceof String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ serverCallbackName_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
+ }
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public Builder setServerCallbackName(
+ java.lang.String value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ serverCallbackName_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public Builder clearServerCallbackName() {
+ bitField0_ = (bitField0_ & ~0x00000001);
+ serverCallbackName_ = getDefaultInstance().getServerCallbackName();
+ onChanged();
+ return this;
+ }
+ /**
+ * required string server_callback_name = 1;
+ */
+ public Builder setServerCallbackNameBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ serverCallbackName_ = value;
+ onChanged();
+ return this;
+ }
+
+ // optional bytes server_callback_initData = 2;
+ private com.google.protobuf.ByteString serverCallbackInitData_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ public boolean hasServerCallbackInitData() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ public com.google.protobuf.ByteString getServerCallbackInitData() {
+ return serverCallbackInitData_;
+ }
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ public Builder setServerCallbackInitData(com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000002;
+ serverCallbackInitData_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * optional bytes server_callback_initData = 2;
+ */
+ public Builder clearServerCallbackInitData() {
+ bitField0_ = (bitField0_ & ~0x00000002);
+ serverCallbackInitData_ = getDefaultInstance().getServerCallbackInitData();
+ onChanged();
+ return this;
+ }
+
+ // repeated .BatchCoprocessorServiceCall call = 3;
+ private java.util.Listrepeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public java.util.Listrepeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public int getCallCount() {
+ if (callBuilder_ == null) {
+ return call_.size();
+ } else {
+ return callBuilder_.getCount();
+ }
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall getCall(int index) {
+ if (callBuilder_ == null) {
+ return call_.get(index);
+ } else {
+ return callBuilder_.getMessage(index);
+ }
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder setCall(
+ int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall value) {
+ if (callBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureCallIsMutable();
+ call_.set(index, value);
+ onChanged();
+ } else {
+ callBuilder_.setMessage(index, value);
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder setCall(
+ int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder builderForValue) {
+ if (callBuilder_ == null) {
+ ensureCallIsMutable();
+ call_.set(index, builderForValue.build());
+ onChanged();
+ } else {
+ callBuilder_.setMessage(index, builderForValue.build());
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder addCall(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall value) {
+ if (callBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureCallIsMutable();
+ call_.add(value);
+ onChanged();
+ } else {
+ callBuilder_.addMessage(value);
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder addCall(
+ int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall value) {
+ if (callBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ ensureCallIsMutable();
+ call_.add(index, value);
+ onChanged();
+ } else {
+ callBuilder_.addMessage(index, value);
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder addCall(
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder builderForValue) {
+ if (callBuilder_ == null) {
+ ensureCallIsMutable();
+ call_.add(builderForValue.build());
+ onChanged();
+ } else {
+ callBuilder_.addMessage(builderForValue.build());
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder addCall(
+ int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder builderForValue) {
+ if (callBuilder_ == null) {
+ ensureCallIsMutable();
+ call_.add(index, builderForValue.build());
+ onChanged();
+ } else {
+ callBuilder_.addMessage(index, builderForValue.build());
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder addAllCall(
+ java.lang.Iterable extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall> values) {
+ if (callBuilder_ == null) {
+ ensureCallIsMutable();
+ super.addAll(values, call_);
+ onChanged();
+ } else {
+ callBuilder_.addAllMessages(values);
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder clearCall() {
+ if (callBuilder_ == null) {
+ call_ = java.util.Collections.emptyList();
+ bitField0_ = (bitField0_ & ~0x00000004);
+ onChanged();
+ } else {
+ callBuilder_.clear();
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public Builder removeCall(int index) {
+ if (callBuilder_ == null) {
+ ensureCallIsMutable();
+ call_.remove(index);
+ onChanged();
+ } else {
+ callBuilder_.remove(index);
+ }
+ return this;
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder getCallBuilder(
+ int index) {
+ return getCallFieldBuilder().getBuilder(index);
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCallOrBuilder getCallOrBuilder(
+ int index) {
+ if (callBuilder_ == null) {
+ return call_.get(index); } else {
+ return callBuilder_.getMessageOrBuilder(index);
+ }
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public java.util.List extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCallOrBuilder>
+ getCallOrBuilderList() {
+ if (callBuilder_ != null) {
+ return callBuilder_.getMessageOrBuilderList();
+ } else {
+ return java.util.Collections.unmodifiableList(call_);
+ }
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder addCallBuilder() {
+ return getCallFieldBuilder().addBuilder(
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.getDefaultInstance());
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.Builder addCallBuilder(
+ int index) {
+ return getCallFieldBuilder().addBuilder(
+ index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceCall.getDefaultInstance());
+ }
+ /**
+ * repeated .BatchCoprocessorServiceCall call = 3;
+ */
+ public java.util.Listrequired .NameBytesPair value = 1;
+ */
+ boolean hasValue();
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair getValue();
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getValueOrBuilder();
+ }
+ /**
+ * Protobuf type {@code BatchCoprocessorServiceResponse}
+ */
+ public static final class BatchCoprocessorServiceResponse extends
+ com.google.protobuf.GeneratedMessage
+ implements BatchCoprocessorServiceResponseOrBuilder {
+ // Use BatchCoprocessorServiceResponse.newBuilder() to construct.
+ private BatchCoprocessorServiceResponse(com.google.protobuf.GeneratedMessage.Builder> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private BatchCoprocessorServiceResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final BatchCoprocessorServiceResponse defaultInstance;
+ public static BatchCoprocessorServiceResponse getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public BatchCoprocessorServiceResponse getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private BatchCoprocessorServiceResponse(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.Builder subBuilder = null;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ subBuilder = value_.toBuilder();
+ }
+ value_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.PARSER, extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(value_);
+ value_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000001;
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_BatchCoprocessorServiceResponse_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_BatchCoprocessorServiceResponse_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse.Builder.class);
+ }
+
+ public static com.google.protobuf.Parserrequired .NameBytesPair value = 1;
+ */
+ public boolean hasValue() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair getValue() {
+ return value_;
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getValueOrBuilder() {
+ return value_;
+ }
+
+ private void initFields() {
+ value_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.getDefaultInstance();
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasValue()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!getValue().isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeMessage(1, value_);
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(1, value_);
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse) obj;
+
+ boolean result = true;
+ result = result && (hasValue() == other.hasValue());
+ if (hasValue()) {
+ result = result && getValue()
+ .equals(other.getValue());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasValue()) {
+ hash = (37 * hash) + VALUE_FIELD_NUMBER;
+ hash = (53 * hash) + getValue().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceResponse prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code BatchCoprocessorServiceResponse}
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builderrequired .NameBytesPair value = 1;
+ */
+ public boolean hasValue() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair getValue() {
+ if (valueBuilder_ == null) {
+ return value_;
+ } else {
+ return valueBuilder_.getMessage();
+ }
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public Builder setValue(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair value) {
+ if (valueBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ value_ = value;
+ onChanged();
+ } else {
+ valueBuilder_.setMessage(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public Builder setValue(
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.Builder builderForValue) {
+ if (valueBuilder_ == null) {
+ value_ = builderForValue.build();
+ onChanged();
+ } else {
+ valueBuilder_.setMessage(builderForValue.build());
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public Builder mergeValue(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair value) {
+ if (valueBuilder_ == null) {
+ if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ value_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.getDefaultInstance()) {
+ value_ =
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.newBuilder(value_).mergeFrom(value).buildPartial();
+ } else {
+ value_ = value;
+ }
+ onChanged();
+ } else {
+ valueBuilder_.mergeFrom(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public Builder clearValue() {
+ if (valueBuilder_ == null) {
+ value_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.getDefaultInstance();
+ onChanged();
+ } else {
+ valueBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000001);
+ return this;
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.Builder getValueBuilder() {
+ bitField0_ |= 0x00000001;
+ onChanged();
+ return getValueFieldBuilder().getBuilder();
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getValueOrBuilder() {
+ if (valueBuilder_ != null) {
+ return valueBuilder_.getMessageOrBuilder();
+ } else {
+ return value_;
+ }
+ }
+ /**
+ * required .NameBytesPair value = 1;
+ */
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder>
+ getValueFieldBuilder() {
+ if (valueBuilder_ == null) {
+ valueBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder>(
+ value_,
+ getParentForChildren(),
+ isClean());
+ value_ = null;
+ }
+ return valueBuilder_;
+ }
+
+ // @@protoc_insertion_point(builder_scope:BatchCoprocessorServiceResponse)
+ }
+
+ static {
+ defaultInstance = new BatchCoprocessorServiceResponse(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:BatchCoprocessorServiceResponse)
+ }
+
+ /**
* Protobuf service {@code ClientService}
*/
public static abstract class ClientService
@@ -28371,6 +30700,14 @@
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiRequest request,
com.google.protobuf.RpcCallbackrpc ExecBatchService(.BatchCoprocessorServiceRequest) returns (.BatchCoprocessorServiceResponse);
+ */
+ public abstract void execBatchService(
+ com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest request,
+ com.google.protobuf.RpcCallbackrpc ExecBatchService(.BatchCoprocessorServiceRequest) returns (.BatchCoprocessorServiceResponse);
+ */
+ public abstract void execBatchService(
+ com.google.protobuf.RpcController controller,
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BatchCoprocessorServiceRequest request,
+ com.google.protobuf.RpcCallbackrow belongs.
+ * @param ServerName
+ * The server name we want to connect.
+ */
+ public BatchRegionServerCallable(HConnection connection, TableName tableName,
+ ServerName serverName) {
+ this.connection = connection;
+ this.tableName = tableName;
+ this.serverName = serverName;
+ }
+
+ /**
+ * Prepare for connection to the server.
+ * @param reload this is not used in this implementation
+ * @throws IOException e
+ */
+ @Override
+ public void prepare(boolean reload) throws IOException {
+ setStub(connection.getClient(serverName));
+ }
+
+ @Override
+ public void throwable(Throwable t, boolean retrying) {
+ if (t instanceof SocketTimeoutException || t instanceof ConnectException
+ || t instanceof RetriesExhaustedException || connection.isDeadServer(serverName)) {
+ // if thrown these exceptions, we clear all the cache entries that
+ // map to that slow/dead server; otherwise, let cache miss and ask
+ // hbase:meta again to find the new location
+ connection.clearCaches(serverName);
+ }
+ }
+
+ @Override
+ public String getExceptionMessageAdditionalDetail() {
+ return "table '" + tableName + " in the server " + serverName;
+ }
+
+ @Override
+ public long sleep(long pause, int tries) {
+ // Tries hasn't been bumped up yet so we use "tries + 1" to get right pause time
+ long sleep = ConnectionUtils.getPauseTime(pause, tries + 1);
+ if (sleep < MIN_WAIT_DEAD_SERVER && connection.isDeadServer(serverName)) {
+ sleep = ConnectionUtils.addJitter(MIN_WAIT_DEAD_SERVER, 0.10f);
+ }
+ return sleep;
+ }
+
+ protected ClientService.BlockingInterface getStub() {
+ return this.stub;
+ }
+
+ void setStub(final ClientService.BlockingInterface stub) {
+ this.stub = stub;
+ }
+
+ public TableName getTableName() {
+ return this.tableName;
+ }
+
+ public ServerName getServerName() {
+ return this.serverName;
+ }
+
+}
Index: hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java
===================================================================
--- hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java (revision 23090)
+++ hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java (revision 23094)
@@ -22,6 +22,7 @@
import java.io.IOException;
import org.apache.hadoop.classification.InterfaceAudience;
+import org.apache.hadoop.hbase.ServerName;
/**
@@ -69,4 +70,23 @@
public interface Callback+ * When used with + * {@link org.apache.hadoop.hbase.client.coprocessor.BatchCoprocessorServiceClient# + * batchCoprocessorService(Class, HTable, byte[], Batch.Call, String, byte[], Batch.BatchCallback)} + * the implementation's {@link Batch.BatchCallback#update(ServerName, Object)} method will be + * called with the {@link Batch.Call#call(Object)} return value from each region server in the + * selected range. + *
+ * + * @param