diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java index 6cabccd..dd50668 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionLoad.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hbase; import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Strings; @@ -31,9 +31,9 @@ import org.apache.hadoop.hbase.util.Strings; @InterfaceAudience.Private public class RegionLoad { - protected HBaseProtos.RegionLoad regionLoadPB; + protected ClusterStatusProtos.RegionLoad regionLoadPB; - public RegionLoad(HBaseProtos.RegionLoad regionLoadPB) { + public RegionLoad(ClusterStatusProtos.RegionLoad regionLoadPB) { this.regionLoadPB = regionLoadPB; } @@ -203,4 +203,4 @@ public class RegionLoad { compactionProgressPct); return sb.toString(); } -} +} \ No newline at end of file diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerLoad.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerLoad.java index 18cdaa0..6f5051b 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerLoad.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerLoad.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor; import org.apache.hadoop.hbase.util.Bytes; @@ -53,9 +54,9 @@ public class ServerLoad { private long totalCompactingKVs = 0; private long currentCompactedKVs = 0; - public ServerLoad(HBaseProtos.ServerLoad serverLoad) { + public ServerLoad(ClusterStatusProtos.ServerLoad serverLoad) { this.serverLoad = serverLoad; - for (HBaseProtos.RegionLoad rl: serverLoad.getRegionLoadsList()) { + for (ClusterStatusProtos.RegionLoad rl: serverLoad.getRegionLoadsList()) { stores += rl.getStores(); storefiles += rl.getStorefiles(); storeUncompressedSizeMB += rl.getStoreUncompressedSizeMB(); @@ -76,11 +77,11 @@ public class ServerLoad { // NOTE: Function name cannot start with "get" because then an OpenDataException is thrown because // HBaseProtos.ServerLoad cannot be converted to an open data type(see HBASE-5967). /* @return the underlying ServerLoad protobuf object */ - public HBaseProtos.ServerLoad obtainServerLoadPB() { + public ClusterStatusProtos.ServerLoad obtainServerLoadPB() { return serverLoad; } - protected HBaseProtos.ServerLoad serverLoad; + protected ClusterStatusProtos.ServerLoad serverLoad; /* @return number of requests since last report. */ public int getNumberOfRequests() { @@ -200,7 +201,7 @@ public class ServerLoad { public Map getRegionsLoad() { Map regionLoads = new TreeMap(Bytes.BYTES_COMPARATOR); - for (HBaseProtos.RegionLoad rl : serverLoad.getRegionLoadsList()) { + for (ClusterStatusProtos.RegionLoad rl : serverLoad.getRegionLoadsList()) { RegionLoad regionLoad = new RegionLoad(rl); regionLoads.put(regionLoad.getName(), regionLoad); } @@ -299,5 +300,5 @@ public class ServerLoad { } public static final ServerLoad EMPTY_SERVERLOAD = - new ServerLoad(HBaseProtos.ServerLoad.newBuilder().build()); + new ServerLoad(ClusterStatusProtos.ServerLoad.newBuilder().build()); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java index b08101c..c9e34d0 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java @@ -24,7 +24,7 @@ import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer; +import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer; import org.apache.hadoop.hbase.util.Addressing; import org.apache.hadoop.hbase.util.Bytes; @@ -339,8 +339,8 @@ public class ServerName implements Comparable { if (ProtobufUtil.isPBMagicPrefix(data)) { int prefixLen = ProtobufUtil.lengthOfPBMagic(); try { - RootRegionServer rss = - RootRegionServer.newBuilder().mergeFrom(data, prefixLen, data.length - prefixLen).build(); + MetaRegionServer rss = + MetaRegionServer.newBuilder().mergeFrom(data, prefixLen, data.length - prefixLen).build(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName sn = rss.getServer(); return new ServerName(sn.getHostName(), sn.getPort(), sn.getStartCode()); } catch (InvalidProtocolBufferException e) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java index b413146..f1ab7883 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java @@ -25,13 +25,10 @@ import java.util.List; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import com.google.protobuf.InvalidProtocolBufferException; @@ -55,7 +52,6 @@ public class FilterList extends Filter { MUST_PASS_ONE } - private static final Configuration conf = HBaseConfiguration.create(); private static final int MAX_LOG_FILTERS = 5; private Operator operator = Operator.MUST_PASS_ALL; private List filters = new ArrayList(); @@ -306,7 +302,7 @@ public class FilterList extends Filter { List rowFilters = new ArrayList(proto.getFiltersCount()); try { - for (HBaseProtos.Filter filter : proto.getFiltersList()) { + for (FilterProtos.Filter filter : proto.getFiltersList()) { rowFilters.add(ProtobufUtil.toFilter(filter)); } } catch (IOException ioe) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 3eb5f94..75d41dd 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -80,6 +80,7 @@ import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.SplitRegionRequest; import org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos; +import org.apache.hadoop.hbase.protobuf.generated.CellProtos; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileResponse; @@ -97,10 +98,11 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Del import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanRequest; import org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos; +import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; import org.apache.hadoop.hbase.protobuf.generated.MapReduceProtos; @@ -383,7 +385,7 @@ public final class ProtobufUtil { get.setTimeRange(minStamp, maxStamp); } if (proto.hasFilter()) { - HBaseProtos.Filter filter = proto.getFilter(); + FilterProtos.Filter filter = proto.getFilter(); get.setFilter(ProtobufUtil.toFilter(filter)); } for (NameBytesPair attribute: proto.getAttributeList()) { @@ -822,7 +824,7 @@ public final class ProtobufUtil { scan.setTimeRange(minStamp, maxStamp); } if (proto.hasFilter()) { - HBaseProtos.Filter filter = proto.getFilter(); + FilterProtos.Filter filter = proto.getFilter(); scan.setFilter(ProtobufUtil.toFilter(filter)); } if (proto.hasBatchSize()) { @@ -1058,9 +1060,9 @@ public final class ProtobufUtil { * @return the converted client Result */ public static Result toResult(final ClientProtos.Result proto) { - List values = proto.getCellList(); + List values = proto.getCellList(); List cells = new ArrayList(values.size()); - for (HBaseProtos.Cell c: values) { + for (CellProtos.Cell c: values) { cells.add(toCell(c)); } return new Result(cells); @@ -1086,9 +1088,9 @@ public final class ProtobufUtil { cells.add(scanner.current()); } } - List values = proto.getCellList(); + List values = proto.getCellList(); if (cells == null) cells = new ArrayList(values.size()); - for (HBaseProtos.Cell c: values) { + for (CellProtos.Cell c: values) { cells.add(toCell(c)); } return new Result(cells); @@ -1139,7 +1141,7 @@ public final class ProtobufUtil { * @return the converted Filter */ @SuppressWarnings("unchecked") - public static Filter toFilter(HBaseProtos.Filter proto) throws IOException { + public static Filter toFilter(FilterProtos.Filter proto) throws IOException { String type = proto.getName(); final byte [] value = proto.getSerializedFilter().toByteArray(); String funcName = "parseFrom"; @@ -1162,8 +1164,8 @@ public final class ProtobufUtil { * @param filter the Filter to convert * @return the converted protocol buffer Filter */ - public static HBaseProtos.Filter toFilter(Filter filter) throws IOException { - HBaseProtos.Filter.Builder builder = HBaseProtos.Filter.newBuilder(); + public static FilterProtos.Filter toFilter(Filter filter) throws IOException { + FilterProtos.Filter.Builder builder = FilterProtos.Filter.newBuilder(); builder.setName(filter.getClass().getName()); builder.setSerializedFilter(ByteString.copyFrom(filter.toByteArray())); return builder.build(); @@ -1960,23 +1962,23 @@ public final class ProtobufUtil { throw new IOException(se); } - public static HBaseProtos.Cell toCell(final Cell kv) { + public static CellProtos.Cell toCell(final Cell kv) { // Doing this is going to kill us if we do it for all data passed. // St.Ack 20121205 - HBaseProtos.Cell.Builder kvbuilder = HBaseProtos.Cell.newBuilder(); + CellProtos.Cell.Builder kvbuilder = CellProtos.Cell.newBuilder(); kvbuilder.setRow(ByteString.copyFrom(kv.getRowArray(), kv.getRowOffset(), kv.getRowLength())); kvbuilder.setFamily(ByteString.copyFrom(kv.getFamilyArray(), kv.getFamilyOffset(), kv.getFamilyLength())); kvbuilder.setQualifier(ByteString.copyFrom(kv.getQualifierArray(), kv.getQualifierOffset(), kv.getQualifierLength())); - kvbuilder.setCellType(HBaseProtos.CellType.valueOf(kv.getTypeByte())); + kvbuilder.setCellType(CellProtos.CellType.valueOf(kv.getTypeByte())); kvbuilder.setTimestamp(kv.getTimestamp()); kvbuilder.setValue(ByteString.copyFrom(kv.getValueArray(), kv.getValueOffset(), kv.getValueLength())); return kvbuilder.build(); } - public static Cell toCell(final HBaseProtos.Cell cell) { + public static Cell toCell(final CellProtos.Cell cell) { // Doing this is going to kill us if we do it for all data passed. // St.Ack 20121205 return CellUtil.createCell(cell.getRow().toByteArray(), diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaRegionTracker.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaRegionTracker.java index 70b731e..aca2708 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaRegionTracker.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaRegionTracker.java @@ -141,8 +141,8 @@ public class MetaRegionTracker extends ZooKeeperNodeTracker { HBaseProtos.ServerName pbsn = HBaseProtos.ServerName.newBuilder().setHostName(sn.getHostname()). setPort(sn.getPort()).setStartCode(sn.getStartcode()).build(); - ZooKeeperProtos.RootRegionServer pbrsr = - ZooKeeperProtos.RootRegionServer.newBuilder().setServer(pbsn).build(); + ZooKeeperProtos.MetaRegionServer pbrsr = + ZooKeeperProtos.MetaRegionServer.newBuilder().setServer(pbsn).build(); return ProtobufUtil.prependPBMagic(pbrsr.toByteArray()); } diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java new file mode 100644 index 0000000..4408081 --- /dev/null +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java @@ -0,0 +1,1622 @@ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: Cell.proto + +package org.apache.hadoop.hbase.protobuf.generated; + +public final class CellProtos { + private CellProtos() {} + public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistry registry) { + } + public enum CellType + implements com.google.protobuf.ProtocolMessageEnum { + MINIMUM(0, 0), + PUT(1, 4), + DELETE(2, 8), + DELETE_COLUMN(3, 12), + DELETE_FAMILY(4, 14), + MAXIMUM(5, 255), + ; + + public static final int MINIMUM_VALUE = 0; + public static final int PUT_VALUE = 4; + public static final int DELETE_VALUE = 8; + public static final int DELETE_COLUMN_VALUE = 12; + public static final int DELETE_FAMILY_VALUE = 14; + public static final int MAXIMUM_VALUE = 255; + + + public final int getNumber() { return value; } + + public static CellType valueOf(int value) { + switch (value) { + case 0: return MINIMUM; + case 4: return PUT; + case 8: return DELETE; + case 12: return DELETE_COLUMN; + case 14: return DELETE_FAMILY; + case 255: return MAXIMUM; + default: return null; + } + } + + public static com.google.protobuf.Internal.EnumLiteMap + internalGetValueMap() { + return internalValueMap; + } + private static com.google.protobuf.Internal.EnumLiteMap + internalValueMap = + new com.google.protobuf.Internal.EnumLiteMap() { + public CellType findValueByNumber(int number) { + return CellType.valueOf(number); + } + }; + + public final com.google.protobuf.Descriptors.EnumValueDescriptor + getValueDescriptor() { + return getDescriptor().getValues().get(index); + } + public final com.google.protobuf.Descriptors.EnumDescriptor + getDescriptorForType() { + return getDescriptor(); + } + public static final com.google.protobuf.Descriptors.EnumDescriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.getDescriptor().getEnumTypes().get(0); + } + + private static final CellType[] VALUES = { + MINIMUM, PUT, DELETE, DELETE_COLUMN, DELETE_FAMILY, MAXIMUM, + }; + + public static CellType valueOf( + com.google.protobuf.Descriptors.EnumValueDescriptor desc) { + if (desc.getType() != getDescriptor()) { + throw new java.lang.IllegalArgumentException( + "EnumValueDescriptor is not for this type."); + } + return VALUES[desc.getIndex()]; + } + + private final int index; + private final int value; + + private CellType(int index, int value) { + this.index = index; + this.value = value; + } + + // @@protoc_insertion_point(enum_scope:CellType) + } + + public interface CellOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // optional bytes row = 1; + boolean hasRow(); + com.google.protobuf.ByteString getRow(); + + // optional bytes family = 2; + boolean hasFamily(); + com.google.protobuf.ByteString getFamily(); + + // optional bytes qualifier = 3; + boolean hasQualifier(); + com.google.protobuf.ByteString getQualifier(); + + // optional uint64 timestamp = 4; + boolean hasTimestamp(); + long getTimestamp(); + + // optional .CellType cellType = 5; + boolean hasCellType(); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType(); + + // optional bytes value = 6; + boolean hasValue(); + com.google.protobuf.ByteString getValue(); + } + public static final class Cell extends + com.google.protobuf.GeneratedMessage + implements CellOrBuilder { + // Use Cell.newBuilder() to construct. + private Cell(Builder builder) { + super(builder); + } + private Cell(boolean noInit) {} + + private static final Cell defaultInstance; + public static Cell getDefaultInstance() { + return defaultInstance; + } + + public Cell getDefaultInstanceForType() { + return defaultInstance; + } + + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_Cell_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_Cell_fieldAccessorTable; + } + + private int bitField0_; + // optional bytes row = 1; + public static final int ROW_FIELD_NUMBER = 1; + private com.google.protobuf.ByteString row_; + public boolean hasRow() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public com.google.protobuf.ByteString getRow() { + return row_; + } + + // optional bytes family = 2; + public static final int FAMILY_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString family_; + public boolean hasFamily() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public com.google.protobuf.ByteString getFamily() { + return family_; + } + + // optional bytes qualifier = 3; + public static final int QUALIFIER_FIELD_NUMBER = 3; + private com.google.protobuf.ByteString qualifier_; + public boolean hasQualifier() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public com.google.protobuf.ByteString getQualifier() { + return qualifier_; + } + + // optional uint64 timestamp = 4; + public static final int TIMESTAMP_FIELD_NUMBER = 4; + private long timestamp_; + public boolean hasTimestamp() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public long getTimestamp() { + return timestamp_; + } + + // optional .CellType cellType = 5; + public static final int CELLTYPE_FIELD_NUMBER = 5; + private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_; + public boolean hasCellType() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType() { + return cellType_; + } + + // optional bytes value = 6; + public static final int VALUE_FIELD_NUMBER = 6; + private com.google.protobuf.ByteString value_; + public boolean hasValue() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public com.google.protobuf.ByteString getValue() { + return value_; + } + + private void initFields() { + row_ = com.google.protobuf.ByteString.EMPTY; + family_ = com.google.protobuf.ByteString.EMPTY; + qualifier_ = com.google.protobuf.ByteString.EMPTY; + timestamp_ = 0L; + cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + value_ = com.google.protobuf.ByteString.EMPTY; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, row_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, family_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeBytes(3, qualifier_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeUInt64(4, timestamp_); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + output.writeEnum(5, cellType_.getNumber()); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + output.writeBytes(6, value_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, row_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, family_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(3, qualifier_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(4, timestamp_); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + size += com.google.protobuf.CodedOutputStream + .computeEnumSize(5, cellType_.getNumber()); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(6, value_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell other = (org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell) obj; + + boolean result = true; + result = result && (hasRow() == other.hasRow()); + if (hasRow()) { + result = result && getRow() + .equals(other.getRow()); + } + result = result && (hasFamily() == other.hasFamily()); + if (hasFamily()) { + result = result && getFamily() + .equals(other.getFamily()); + } + result = result && (hasQualifier() == other.hasQualifier()); + if (hasQualifier()) { + result = result && getQualifier() + .equals(other.getQualifier()); + } + result = result && (hasTimestamp() == other.hasTimestamp()); + if (hasTimestamp()) { + result = result && (getTimestamp() + == other.getTimestamp()); + } + result = result && (hasCellType() == other.hasCellType()); + if (hasCellType()) { + result = result && + (getCellType() == other.getCellType()); + } + result = result && (hasValue() == other.hasValue()); + if (hasValue()) { + result = result && getValue() + .equals(other.getValue()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + @java.lang.Override + public int hashCode() { + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasRow()) { + hash = (37 * hash) + ROW_FIELD_NUMBER; + hash = (53 * hash) + getRow().hashCode(); + } + if (hasFamily()) { + hash = (37 * hash) + FAMILY_FIELD_NUMBER; + hash = (53 * hash) + getFamily().hashCode(); + } + if (hasQualifier()) { + hash = (37 * hash) + QUALIFIER_FIELD_NUMBER; + hash = (53 * hash) + getQualifier().hashCode(); + } + if (hasTimestamp()) { + hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getTimestamp()); + } + if (hasCellType()) { + hash = (37 * hash) + CELLTYPE_FIELD_NUMBER; + hash = (53 * hash) + hashEnum(getCellType()); + } + if (hasValue()) { + hash = (37 * hash) + VALUE_FIELD_NUMBER; + hash = (53 * hash) + getValue().hashCode(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom(java.io.InputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input, extensionRegistry)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_Cell_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_Cell_fieldAccessorTable; + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + row_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + family_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000002); + qualifier_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000004); + timestamp_ = 0L; + bitField0_ = (bitField0_ & ~0x00000008); + cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + bitField0_ = (bitField0_ & ~0x00000010); + value_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000020); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.getDescriptor(); + } + + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell build() { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + private org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell buildParsed() + throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException( + result).asInvalidProtocolBufferException(); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell result = new org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.row_ = row_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.family_ = family_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.qualifier_ = qualifier_; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000008; + } + result.timestamp_ = timestamp_; + if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + to_bitField0_ |= 0x00000010; + } + result.cellType_ = cellType_; + if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + to_bitField0_ |= 0x00000020; + } + result.value_ = value_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.getDefaultInstance()) return this; + if (other.hasRow()) { + setRow(other.getRow()); + } + if (other.hasFamily()) { + setFamily(other.getFamily()); + } + if (other.hasQualifier()) { + setQualifier(other.getQualifier()); + } + if (other.hasTimestamp()) { + setTimestamp(other.getTimestamp()); + } + if (other.hasCellType()) { + setCellType(other.getCellType()); + } + if (other.hasValue()) { + setValue(other.getValue()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder( + this.getUnknownFields()); + while (true) { + int tag = input.readTag(); + switch (tag) { + case 0: + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + row_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + family_ = input.readBytes(); + break; + } + case 26: { + bitField0_ |= 0x00000004; + qualifier_ = input.readBytes(); + break; + } + case 32: { + bitField0_ |= 0x00000008; + timestamp_ = input.readUInt64(); + break; + } + case 40: { + int rawValue = input.readEnum(); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType value = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.valueOf(rawValue); + if (value == null) { + unknownFields.mergeVarintField(5, rawValue); + } else { + bitField0_ |= 0x00000010; + cellType_ = value; + } + break; + } + case 50: { + bitField0_ |= 0x00000020; + value_ = input.readBytes(); + break; + } + } + } + } + + private int bitField0_; + + // optional bytes row = 1; + private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasRow() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public com.google.protobuf.ByteString getRow() { + return row_; + } + public Builder setRow(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + row_ = value; + onChanged(); + return this; + } + public Builder clearRow() { + bitField0_ = (bitField0_ & ~0x00000001); + row_ = getDefaultInstance().getRow(); + onChanged(); + return this; + } + + // optional bytes family = 2; + private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasFamily() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public com.google.protobuf.ByteString getFamily() { + return family_; + } + public Builder setFamily(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + family_ = value; + onChanged(); + return this; + } + public Builder clearFamily() { + bitField0_ = (bitField0_ & ~0x00000002); + family_ = getDefaultInstance().getFamily(); + onChanged(); + return this; + } + + // optional bytes qualifier = 3; + private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasQualifier() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public com.google.protobuf.ByteString getQualifier() { + return qualifier_; + } + public Builder setQualifier(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000004; + qualifier_ = value; + onChanged(); + return this; + } + public Builder clearQualifier() { + bitField0_ = (bitField0_ & ~0x00000004); + qualifier_ = getDefaultInstance().getQualifier(); + onChanged(); + return this; + } + + // optional uint64 timestamp = 4; + private long timestamp_ ; + public boolean hasTimestamp() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public long getTimestamp() { + return timestamp_; + } + public Builder setTimestamp(long value) { + bitField0_ |= 0x00000008; + timestamp_ = value; + onChanged(); + return this; + } + public Builder clearTimestamp() { + bitField0_ = (bitField0_ & ~0x00000008); + timestamp_ = 0L; + onChanged(); + return this; + } + + // optional .CellType cellType = 5; + private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + public boolean hasCellType() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType() { + return cellType_; + } + public Builder setCellType(org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000010; + cellType_ = value; + onChanged(); + return this; + } + public Builder clearCellType() { + bitField0_ = (bitField0_ & ~0x00000010); + cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + onChanged(); + return this; + } + + // optional bytes value = 6; + private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasValue() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public com.google.protobuf.ByteString getValue() { + return value_; + } + public Builder setValue(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000020; + value_ = value; + onChanged(); + return this; + } + public Builder clearValue() { + bitField0_ = (bitField0_ & ~0x00000020); + value_ = getDefaultInstance().getValue(); + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:Cell) + } + + static { + defaultInstance = new Cell(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:Cell) + } + + public interface KeyValueOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // required bytes row = 1; + boolean hasRow(); + com.google.protobuf.ByteString getRow(); + + // required bytes family = 2; + boolean hasFamily(); + com.google.protobuf.ByteString getFamily(); + + // required bytes qualifier = 3; + boolean hasQualifier(); + com.google.protobuf.ByteString getQualifier(); + + // optional uint64 timestamp = 4; + boolean hasTimestamp(); + long getTimestamp(); + + // optional .CellType keyType = 5; + boolean hasKeyType(); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType(); + + // optional bytes value = 6; + boolean hasValue(); + com.google.protobuf.ByteString getValue(); + } + public static final class KeyValue extends + com.google.protobuf.GeneratedMessage + implements KeyValueOrBuilder { + // Use KeyValue.newBuilder() to construct. + private KeyValue(Builder builder) { + super(builder); + } + private KeyValue(boolean noInit) {} + + private static final KeyValue defaultInstance; + public static KeyValue getDefaultInstance() { + return defaultInstance; + } + + public KeyValue getDefaultInstanceForType() { + return defaultInstance; + } + + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_KeyValue_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_KeyValue_fieldAccessorTable; + } + + private int bitField0_; + // required bytes row = 1; + public static final int ROW_FIELD_NUMBER = 1; + private com.google.protobuf.ByteString row_; + public boolean hasRow() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public com.google.protobuf.ByteString getRow() { + return row_; + } + + // required bytes family = 2; + public static final int FAMILY_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString family_; + public boolean hasFamily() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public com.google.protobuf.ByteString getFamily() { + return family_; + } + + // required bytes qualifier = 3; + public static final int QUALIFIER_FIELD_NUMBER = 3; + private com.google.protobuf.ByteString qualifier_; + public boolean hasQualifier() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public com.google.protobuf.ByteString getQualifier() { + return qualifier_; + } + + // optional uint64 timestamp = 4; + public static final int TIMESTAMP_FIELD_NUMBER = 4; + private long timestamp_; + public boolean hasTimestamp() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public long getTimestamp() { + return timestamp_; + } + + // optional .CellType keyType = 5; + public static final int KEYTYPE_FIELD_NUMBER = 5; + private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_; + public boolean hasKeyType() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType() { + return keyType_; + } + + // optional bytes value = 6; + public static final int VALUE_FIELD_NUMBER = 6; + private com.google.protobuf.ByteString value_; + public boolean hasValue() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public com.google.protobuf.ByteString getValue() { + return value_; + } + + private void initFields() { + row_ = com.google.protobuf.ByteString.EMPTY; + family_ = com.google.protobuf.ByteString.EMPTY; + qualifier_ = com.google.protobuf.ByteString.EMPTY; + timestamp_ = 0L; + keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + value_ = com.google.protobuf.ByteString.EMPTY; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + if (!hasRow()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasFamily()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasQualifier()) { + memoizedIsInitialized = 0; + return false; + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, row_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, family_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeBytes(3, qualifier_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeUInt64(4, timestamp_); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + output.writeEnum(5, keyType_.getNumber()); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + output.writeBytes(6, value_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, row_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, family_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(3, qualifier_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(4, timestamp_); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + size += com.google.protobuf.CodedOutputStream + .computeEnumSize(5, keyType_.getNumber()); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(6, value_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue other = (org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue) obj; + + boolean result = true; + result = result && (hasRow() == other.hasRow()); + if (hasRow()) { + result = result && getRow() + .equals(other.getRow()); + } + result = result && (hasFamily() == other.hasFamily()); + if (hasFamily()) { + result = result && getFamily() + .equals(other.getFamily()); + } + result = result && (hasQualifier() == other.hasQualifier()); + if (hasQualifier()) { + result = result && getQualifier() + .equals(other.getQualifier()); + } + result = result && (hasTimestamp() == other.hasTimestamp()); + if (hasTimestamp()) { + result = result && (getTimestamp() + == other.getTimestamp()); + } + result = result && (hasKeyType() == other.hasKeyType()); + if (hasKeyType()) { + result = result && + (getKeyType() == other.getKeyType()); + } + result = result && (hasValue() == other.hasValue()); + if (hasValue()) { + result = result && getValue() + .equals(other.getValue()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + @java.lang.Override + public int hashCode() { + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasRow()) { + hash = (37 * hash) + ROW_FIELD_NUMBER; + hash = (53 * hash) + getRow().hashCode(); + } + if (hasFamily()) { + hash = (37 * hash) + FAMILY_FIELD_NUMBER; + hash = (53 * hash) + getFamily().hashCode(); + } + if (hasQualifier()) { + hash = (37 * hash) + QUALIFIER_FIELD_NUMBER; + hash = (53 * hash) + getQualifier().hashCode(); + } + if (hasTimestamp()) { + hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getTimestamp()); + } + if (hasKeyType()) { + hash = (37 * hash) + KEYTYPE_FIELD_NUMBER; + hash = (53 * hash) + hashEnum(getKeyType()); + } + if (hasValue()) { + hash = (37 * hash) + VALUE_FIELD_NUMBER; + hash = (53 * hash) + getValue().hashCode(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom(java.io.InputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input, extensionRegistry)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValueOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_KeyValue_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.internal_static_KeyValue_fieldAccessorTable; + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + row_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + family_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000002); + qualifier_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000004); + timestamp_ = 0L; + bitField0_ = (bitField0_ & ~0x00000008); + keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + bitField0_ = (bitField0_ & ~0x00000010); + value_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000020); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue.getDescriptor(); + } + + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue build() { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + private org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue buildParsed() + throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException( + result).asInvalidProtocolBufferException(); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue result = new org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.row_ = row_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.family_ = family_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.qualifier_ = qualifier_; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000008; + } + result.timestamp_ = timestamp_; + if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + to_bitField0_ |= 0x00000010; + } + result.keyType_ = keyType_; + if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + to_bitField0_ |= 0x00000020; + } + result.value_ = value_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue.getDefaultInstance()) return this; + if (other.hasRow()) { + setRow(other.getRow()); + } + if (other.hasFamily()) { + setFamily(other.getFamily()); + } + if (other.hasQualifier()) { + setQualifier(other.getQualifier()); + } + if (other.hasTimestamp()) { + setTimestamp(other.getTimestamp()); + } + if (other.hasKeyType()) { + setKeyType(other.getKeyType()); + } + if (other.hasValue()) { + setValue(other.getValue()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + if (!hasRow()) { + + return false; + } + if (!hasFamily()) { + + return false; + } + if (!hasQualifier()) { + + return false; + } + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder( + this.getUnknownFields()); + while (true) { + int tag = input.readTag(); + switch (tag) { + case 0: + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + row_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + family_ = input.readBytes(); + break; + } + case 26: { + bitField0_ |= 0x00000004; + qualifier_ = input.readBytes(); + break; + } + case 32: { + bitField0_ |= 0x00000008; + timestamp_ = input.readUInt64(); + break; + } + case 40: { + int rawValue = input.readEnum(); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType value = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.valueOf(rawValue); + if (value == null) { + unknownFields.mergeVarintField(5, rawValue); + } else { + bitField0_ |= 0x00000010; + keyType_ = value; + } + break; + } + case 50: { + bitField0_ |= 0x00000020; + value_ = input.readBytes(); + break; + } + } + } + } + + private int bitField0_; + + // required bytes row = 1; + private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasRow() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public com.google.protobuf.ByteString getRow() { + return row_; + } + public Builder setRow(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + row_ = value; + onChanged(); + return this; + } + public Builder clearRow() { + bitField0_ = (bitField0_ & ~0x00000001); + row_ = getDefaultInstance().getRow(); + onChanged(); + return this; + } + + // required bytes family = 2; + private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasFamily() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public com.google.protobuf.ByteString getFamily() { + return family_; + } + public Builder setFamily(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + family_ = value; + onChanged(); + return this; + } + public Builder clearFamily() { + bitField0_ = (bitField0_ & ~0x00000002); + family_ = getDefaultInstance().getFamily(); + onChanged(); + return this; + } + + // required bytes qualifier = 3; + private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasQualifier() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public com.google.protobuf.ByteString getQualifier() { + return qualifier_; + } + public Builder setQualifier(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000004; + qualifier_ = value; + onChanged(); + return this; + } + public Builder clearQualifier() { + bitField0_ = (bitField0_ & ~0x00000004); + qualifier_ = getDefaultInstance().getQualifier(); + onChanged(); + return this; + } + + // optional uint64 timestamp = 4; + private long timestamp_ ; + public boolean hasTimestamp() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public long getTimestamp() { + return timestamp_; + } + public Builder setTimestamp(long value) { + bitField0_ |= 0x00000008; + timestamp_ = value; + onChanged(); + return this; + } + public Builder clearTimestamp() { + bitField0_ = (bitField0_ & ~0x00000008); + timestamp_ = 0L; + onChanged(); + return this; + } + + // optional .CellType keyType = 5; + private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + public boolean hasKeyType() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType() { + return keyType_; + } + public Builder setKeyType(org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000010; + keyType_ = value; + onChanged(); + return this; + } + public Builder clearKeyType() { + bitField0_ = (bitField0_ & ~0x00000010); + keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; + onChanged(); + return this; + } + + // optional bytes value = 6; + private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasValue() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public com.google.protobuf.ByteString getValue() { + return value_; + } + public Builder setValue(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000020; + value_ = value; + onChanged(); + return this; + } + public Builder clearValue() { + bitField0_ = (bitField0_ & ~0x00000020); + value_ = getDefaultInstance().getValue(); + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:KeyValue) + } + + static { + defaultInstance = new KeyValue(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:KeyValue) + } + + private static com.google.protobuf.Descriptors.Descriptor + internal_static_Cell_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_Cell_fieldAccessorTable; + private static com.google.protobuf.Descriptors.Descriptor + internal_static_KeyValue_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_KeyValue_fieldAccessorTable; + + public static com.google.protobuf.Descriptors.FileDescriptor + getDescriptor() { + return descriptor; + } + private static com.google.protobuf.Descriptors.FileDescriptor + descriptor; + static { + java.lang.String[] descriptorData = { + "\n\nCell.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" + + "ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" + + "p\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n\005" + + "value\030\006 \001(\014\"x\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n\006" + + "family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttimes" + + "tamp\030\004 \001(\004\022\032\n\007keyType\030\005 \001(\0162\t.CellType\022\r" + + "\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000\022\007" + + "\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014\022\021" + + "\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org.a" + + "pache.hadoop.hbase.protobuf.generatedB\nC", + "ellProtosH\001\240\001\001" + }; + com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public com.google.protobuf.ExtensionRegistry assignDescriptors( + com.google.protobuf.Descriptors.FileDescriptor root) { + descriptor = root; + internal_static_Cell_descriptor = + getDescriptor().getMessageTypes().get(0); + internal_static_Cell_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_Cell_descriptor, + new java.lang.String[] { "Row", "Family", "Qualifier", "Timestamp", "CellType", "Value", }, + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.class, + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder.class); + internal_static_KeyValue_descriptor = + getDescriptor().getMessageTypes().get(1); + internal_static_KeyValue_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_KeyValue_descriptor, + new java.lang.String[] { "Row", "Family", "Qualifier", "Timestamp", "KeyType", "Value", }, + org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue.class, + org.apache.hadoop.hbase.protobuf.generated.CellProtos.KeyValue.Builder.class); + return null; + } + }; + com.google.protobuf.Descriptors.FileDescriptor + .internalBuildGeneratedFileFrom(descriptorData, + new com.google.protobuf.Descriptors.FileDescriptor[] { + }, assigner); + } + + // @@protoc_insertion_point(outer_class_scope) +} diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java index 6eb798e..c376e9a 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java @@ -530,8 +530,8 @@ public final class ClientProtos { // optional .Filter filter = 4; boolean hasFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); // optional .TimeRange timeRange = 5; boolean hasTimeRange(); @@ -637,14 +637,14 @@ public final class ClientProtos { // optional .Filter filter = 4; public static final int FILTER_FIELD_NUMBER = 4; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_; + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_; public boolean hasFilter() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { return filter_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { return filter_; } @@ -705,7 +705,7 @@ public final class ClientProtos { row_ = com.google.protobuf.ByteString.EMPTY; column_ = java.util.Collections.emptyList(); attribute_ = java.util.Collections.emptyList(); - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); timeRange_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); maxVersions_ = 1; cacheBlocks_ = true; @@ -1060,7 +1060,7 @@ public final class ClientProtos { attributeBuilder_.clear(); } if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } else { filterBuilder_.clear(); } @@ -1331,7 +1331,7 @@ public final class ClientProtos { break; } case 34: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(); if (hasFilter()) { subBuilder.mergeFrom(getFilter()); } @@ -1771,20 +1771,20 @@ public final class ClientProtos { } // optional .Filter filter = 4; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> filterBuilder_; + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> filterBuilder_; public boolean hasFilter() { return ((bitField0_ & 0x00000008) == 0x00000008); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { if (filterBuilder_ == null) { return filter_; } else { return filterBuilder_.getMessage(); } } - public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -1798,7 +1798,7 @@ public final class ClientProtos { return this; } public Builder setFilter( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filterBuilder_ == null) { filter_ = builderForValue.build(); onChanged(); @@ -1808,12 +1808,12 @@ public final class ClientProtos { bitField0_ |= 0x00000008; return this; } - public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && - filter_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()) { + filter_ != org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()) { filter_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); } else { filter_ = value; } @@ -1826,7 +1826,7 @@ public final class ClientProtos { } public Builder clearFilter() { if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); onChanged(); } else { filterBuilder_.clear(); @@ -1834,12 +1834,12 @@ public final class ClientProtos { bitField0_ = (bitField0_ & ~0x00000008); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder getFilterBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder getFilterBuilder() { bitField0_ |= 0x00000008; onChanged(); return getFilterFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { if (filterBuilder_ != null) { return filterBuilder_.getMessageOrBuilder(); } else { @@ -1847,11 +1847,11 @@ public final class ClientProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> getFilterFieldBuilder() { if (filterBuilder_ == null) { filterBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder>( filter_, getParentForChildren(), isClean()); @@ -2049,13 +2049,13 @@ public final class ClientProtos { extends com.google.protobuf.MessageOrBuilder { // repeated .Cell cell = 1; - java.util.List + java.util.List getCellList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell getCell(int index); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell getCell(int index); int getCellCount(); - java.util.List + java.util.List getCellOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder getCellOrBuilder( + org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder getCellOrBuilder( int index); // optional int32 associatedCellCount = 2; @@ -2093,21 +2093,21 @@ public final class ClientProtos { private int bitField0_; // repeated .Cell cell = 1; public static final int CELL_FIELD_NUMBER = 1; - private java.util.List cell_; - public java.util.List getCellList() { + private java.util.List cell_; + public java.util.List getCellList() { return cell_; } - public java.util.List + public java.util.List getCellOrBuilderList() { return cell_; } public int getCellCount() { return cell_.size(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell getCell(int index) { + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell getCell(int index) { return cell_.get(index); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder getCellOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder getCellOrBuilder( int index) { return cell_.get(index); } @@ -2461,7 +2461,7 @@ public final class ClientProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.newBuilder(); input.readMessage(subBuilder, extensionRegistry); addCell(subBuilder.buildPartial()); break; @@ -2478,19 +2478,19 @@ public final class ClientProtos { private int bitField0_; // repeated .Cell cell = 1; - private java.util.List cell_ = + private java.util.List cell_ = java.util.Collections.emptyList(); private void ensureCellIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - cell_ = new java.util.ArrayList(cell_); + cell_ = new java.util.ArrayList(cell_); bitField0_ |= 0x00000001; } } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder> cellBuilder_; + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder, org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder> cellBuilder_; - public java.util.List getCellList() { + public java.util.List getCellList() { if (cellBuilder_ == null) { return java.util.Collections.unmodifiableList(cell_); } else { @@ -2504,7 +2504,7 @@ public final class ClientProtos { return cellBuilder_.getCount(); } } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell getCell(int index) { + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell getCell(int index) { if (cellBuilder_ == null) { return cell_.get(index); } else { @@ -2512,7 +2512,7 @@ public final class ClientProtos { } } public Builder setCell( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell value) { + int index, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell value) { if (cellBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -2526,7 +2526,7 @@ public final class ClientProtos { return this; } public Builder setCell( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder builderForValue) { if (cellBuilder_ == null) { ensureCellIsMutable(); cell_.set(index, builderForValue.build()); @@ -2536,7 +2536,7 @@ public final class ClientProtos { } return this; } - public Builder addCell(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell value) { + public Builder addCell(org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell value) { if (cellBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -2550,7 +2550,7 @@ public final class ClientProtos { return this; } public Builder addCell( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell value) { + int index, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell value) { if (cellBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -2564,7 +2564,7 @@ public final class ClientProtos { return this; } public Builder addCell( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder builderForValue) { if (cellBuilder_ == null) { ensureCellIsMutable(); cell_.add(builderForValue.build()); @@ -2575,7 +2575,7 @@ public final class ClientProtos { return this; } public Builder addCell( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder builderForValue) { if (cellBuilder_ == null) { ensureCellIsMutable(); cell_.add(index, builderForValue.build()); @@ -2586,7 +2586,7 @@ public final class ClientProtos { return this; } public Builder addAllCell( - java.lang.Iterable values) { + java.lang.Iterable values) { if (cellBuilder_ == null) { ensureCellIsMutable(); super.addAll(values, cell_); @@ -2616,18 +2616,18 @@ public final class ClientProtos { } return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder getCellBuilder( + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder getCellBuilder( int index) { return getCellFieldBuilder().getBuilder(index); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder getCellOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder getCellOrBuilder( int index) { if (cellBuilder_ == null) { return cell_.get(index); } else { return cellBuilder_.getMessageOrBuilder(index); } } - public java.util.List + public java.util.List getCellOrBuilderList() { if (cellBuilder_ != null) { return cellBuilder_.getMessageOrBuilderList(); @@ -2635,25 +2635,25 @@ public final class ClientProtos { return java.util.Collections.unmodifiableList(cell_); } } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder addCellBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder addCellBuilder() { return getCellFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.getDefaultInstance()); + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.getDefaultInstance()); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder addCellBuilder( + public org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder addCellBuilder( int index) { return getCellFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.getDefaultInstance()); + index, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.getDefaultInstance()); } - public java.util.List + public java.util.List getCellBuilderList() { return getCellFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder, org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder> getCellFieldBuilder() { if (cellBuilder_ == null) { cellBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell, org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell.Builder, org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder>( cell_, ((bitField0_ & 0x00000001) == 0x00000001), getParentForChildren(), @@ -10587,8 +10587,8 @@ public final class ClientProtos { // optional .Filter filter = 5; boolean hasFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); // optional .TimeRange timeRange = 6; boolean hasTimeRange(); @@ -10724,14 +10724,14 @@ public final class ClientProtos { // optional .Filter filter = 5; public static final int FILTER_FIELD_NUMBER = 5; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_; + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_; public boolean hasFilter() { return ((bitField0_ & 0x00000004) == 0x00000004); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { return filter_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { return filter_; } @@ -10843,7 +10843,7 @@ public final class ClientProtos { attribute_ = java.util.Collections.emptyList(); startRow_ = com.google.protobuf.ByteString.EMPTY; stopRow_ = com.google.protobuf.ByteString.EMPTY; - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); timeRange_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); maxVersions_ = 1; cacheBlocks_ = true; @@ -11297,7 +11297,7 @@ public final class ClientProtos { stopRow_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000008); if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } else { filterBuilder_.clear(); } @@ -11621,7 +11621,7 @@ public final class ClientProtos { break; } case 42: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(); if (hasFilter()) { subBuilder.mergeFrom(getFilter()); } @@ -12110,20 +12110,20 @@ public final class ClientProtos { } // optional .Filter filter = 5; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> filterBuilder_; + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> filterBuilder_; public boolean hasFilter() { return ((bitField0_ & 0x00000010) == 0x00000010); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { if (filterBuilder_ == null) { return filter_; } else { return filterBuilder_.getMessage(); } } - public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -12137,7 +12137,7 @@ public final class ClientProtos { return this; } public Builder setFilter( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filterBuilder_ == null) { filter_ = builderForValue.build(); onChanged(); @@ -12147,12 +12147,12 @@ public final class ClientProtos { bitField0_ |= 0x00000010; return this; } - public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && - filter_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()) { + filter_ != org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()) { filter_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); } else { filter_ = value; } @@ -12165,7 +12165,7 @@ public final class ClientProtos { } public Builder clearFilter() { if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); onChanged(); } else { filterBuilder_.clear(); @@ -12173,12 +12173,12 @@ public final class ClientProtos { bitField0_ = (bitField0_ & ~0x00000010); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder getFilterBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder getFilterBuilder() { bitField0_ |= 0x00000010; onChanged(); return getFilterFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { if (filterBuilder_ != null) { return filterBuilder_.getMessageOrBuilder(); } else { @@ -12186,11 +12186,11 @@ public final class ClientProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> getFilterFieldBuilder() { if (filterBuilder_ == null) { filterBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder>( filter_, getParentForChildren(), isClean()); @@ -21554,99 +21554,100 @@ public final class ClientProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\014Client.proto\032\013hbase.proto\032\020Comparator." + - "proto\"+\n\006Column\022\016\n\006family\030\001 \002(\014\022\021\n\tquali" + - "fier\030\002 \003(\014\"\342\001\n\003Get\022\013\n\003row\030\001 \002(\014\022\027\n\006colum" + - "n\030\002 \003(\0132\007.Column\022!\n\tattribute\030\003 \003(\0132\016.Na" + - "meBytesPair\022\027\n\006filter\030\004 \001(\0132\007.Filter\022\035\n\t" + - "timeRange\030\005 \001(\0132\n.TimeRange\022\026\n\013maxVersio" + - "ns\030\006 \001(\r:\0011\022\031\n\013cacheBlocks\030\007 \001(\010:\004true\022\022" + - "\n\nstoreLimit\030\010 \001(\r\022\023\n\013storeOffset\030\t \001(\r\"" + - ":\n\006Result\022\023\n\004cell\030\001 \003(\0132\005.Cell\022\033\n\023associ" + - "atedCellCount\030\002 \001(\005\"r\n\nGetRequest\022 \n\006reg", - "ion\030\001 \002(\0132\020.RegionSpecifier\022\021\n\003get\030\002 \002(\013" + - "2\004.Get\022\030\n\020closestRowBefore\030\003 \001(\010\022\025\n\rexis" + - "tenceOnly\030\004 \001(\010\"w\n\017MultiGetRequest\022 \n\006re" + - "gion\030\001 \002(\0132\020.RegionSpecifier\022\021\n\003get\030\002 \003(" + - "\0132\004.Get\022\030\n\020closestRowBefore\030\003 \001(\010\022\025\n\rexi" + - "stenceOnly\030\004 \001(\010\"6\n\013GetResponse\022\027\n\006resul" + - "t\030\001 \001(\0132\007.Result\022\016\n\006exists\030\002 \001(\010\";\n\020Mult" + - "iGetResponse\022\027\n\006result\030\001 \003(\0132\007.Result\022\016\n" + - "\006exists\030\002 \003(\010\"\177\n\tCondition\022\013\n\003row\030\001 \002(\014\022" + - "\016\n\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022!\n\013co", - "mpareType\030\004 \002(\0162\014.CompareType\022\037\n\ncompara" + - "tor\030\005 \002(\0132\013.Comparator\"\365\005\n\rMutationProto" + - "\022\013\n\003row\030\001 \001(\014\022/\n\nmutateType\030\002 \001(\0162\033.Muta" + - "tionProto.MutationType\022/\n\013columnValue\030\003 " + - "\003(\0132\032.MutationProto.ColumnValue\022\021\n\ttimes" + - "tamp\030\004 \001(\004\022!\n\tattribute\030\005 \003(\0132\016.NameByte" + - "sPair\022:\n\ndurability\030\006 \001(\0162\031.MutationProt" + - "o.Durability:\013USE_DEFAULT\022\035\n\ttimeRange\030\007" + - " \001(\0132\n.TimeRange\022\033\n\023associatedCellCount\030" + - "\010 \001(\005\032\326\001\n\013ColumnValue\022\016\n\006family\030\001 \002(\014\022A\n", - "\016qualifierValue\030\002 \003(\0132).MutationProto.Co" + - "lumnValue.QualifierValue\032t\n\016QualifierVal" + - "ue\022\021\n\tqualifier\030\001 \001(\014\022\r\n\005value\030\002 \001(\014\022\021\n\t" + - "timestamp\030\003 \001(\004\022-\n\ndeleteType\030\004 \001(\0162\031.Mu" + - "tationProto.DeleteType\"W\n\nDurability\022\017\n\013" + - "USE_DEFAULT\020\000\022\014\n\010SKIP_WAL\020\001\022\r\n\tASYNC_WAL" + - "\020\002\022\014\n\010SYNC_WAL\020\003\022\r\n\tFSYNC_WAL\020\004\">\n\014Mutat" + - "ionType\022\n\n\006APPEND\020\000\022\r\n\tINCREMENT\020\001\022\007\n\003PU" + - "T\020\002\022\n\n\006DELETE\020\003\"U\n\nDeleteType\022\026\n\022DELETE_" + - "ONE_VERSION\020\000\022\034\n\030DELETE_MULTIPLE_VERSION", - "S\020\001\022\021\n\rDELETE_FAMILY\020\002\"r\n\rMutateRequest\022" + - " \n\006region\030\001 \002(\0132\020.RegionSpecifier\022 \n\010mut" + - "ation\030\002 \002(\0132\016.MutationProto\022\035\n\tcondition" + - "\030\003 \001(\0132\n.Condition\"<\n\016MutateResponse\022\027\n\006" + - "result\030\001 \001(\0132\007.Result\022\021\n\tprocessed\030\002 \001(\010" + - "\"\362\002\n\004Scan\022\027\n\006column\030\001 \003(\0132\007.Column\022!\n\tat" + - "tribute\030\002 \003(\0132\016.NameBytesPair\022\020\n\010startRo" + - "w\030\003 \001(\014\022\017\n\007stopRow\030\004 \001(\014\022\027\n\006filter\030\005 \001(\013" + - "2\007.Filter\022\035\n\ttimeRange\030\006 \001(\0132\n.TimeRange" + - "\022\026\n\013maxVersions\030\007 \001(\r:\0011\022\031\n\013cacheBlocks\030", - "\010 \001(\010:\004true\022\021\n\tbatchSize\030\t \001(\r\022\025\n\rmaxRes" + - "ultSize\030\n \001(\004\022\022\n\nstoreLimit\030\013 \001(\r\022\023\n\013sto" + - "reOffset\030\014 \001(\r\022\"\n\032loadColumnFamiliesOnDe" + - "mand\030\r \001(\010\022\024\n\014cachingCount\030\016 \001(\r\022\023\n\013pref" + - "etching\030\017 \001(\010\"\230\001\n\013ScanRequest\022 \n\006region\030" + - "\001 \001(\0132\020.RegionSpecifier\022\023\n\004scan\030\002 \001(\0132\005." + - "Scan\022\021\n\tscannerId\030\003 \001(\004\022\024\n\014numberOfRows\030" + - "\004 \001(\r\022\024\n\014closeScanner\030\005 \001(\010\022\023\n\013nextCallS" + - "eq\030\006 \001(\004\"l\n\014ScanResponse\022\'\n\016resultCellMe" + - "ta\030\001 \001(\0132\017.ResultCellMeta\022\021\n\tscannerId\030\002", - " \001(\004\022\023\n\013moreResults\030\003 \001(\010\022\013\n\003ttl\030\004 \001(\r\"%" + - "\n\016ResultCellMeta\022\023\n\013cellsLength\030\001 \003(\r\"\260\001" + - "\n\024BulkLoadHFileRequest\022 \n\006region\030\001 \002(\0132\020" + - ".RegionSpecifier\0224\n\nfamilyPath\030\002 \003(\0132 .B" + - "ulkLoadHFileRequest.FamilyPath\022\024\n\014assign" + - "SeqNum\030\003 \001(\010\032*\n\nFamilyPath\022\016\n\006family\030\001 \002" + - "(\014\022\014\n\004path\030\002 \002(\t\"\'\n\025BulkLoadHFileRespons" + - "e\022\016\n\006loaded\030\001 \002(\010\"_\n\026CoprocessorServiceC" + - "all\022\013\n\003row\030\001 \002(\014\022\023\n\013serviceName\030\002 \002(\t\022\022\n" + - "\nmethodName\030\003 \002(\t\022\017\n\007request\030\004 \002(\014\"d\n\031Co", - "processorServiceRequest\022 \n\006region\030\001 \002(\0132" + - "\020.RegionSpecifier\022%\n\004call\030\002 \002(\0132\027.Coproc" + - "essorServiceCall\"]\n\032CoprocessorServiceRe" + - "sponse\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier" + - "\022\035\n\005value\030\002 \002(\0132\016.NameBytesPair\"B\n\013Multi" + - "Action\022 \n\010mutation\030\001 \001(\0132\016.MutationProto" + - "\022\021\n\003get\030\002 \001(\0132\004.Get\"I\n\014ActionResult\022\026\n\005v" + - "alue\030\001 \001(\0132\007.Result\022!\n\texception\030\002 \001(\0132\016" + - ".NameBytesPair\"^\n\014MultiRequest\022 \n\006region" + - "\030\001 \002(\0132\020.RegionSpecifier\022\034\n\006action\030\002 \003(\013", - "2\014.MultiAction\022\016\n\006atomic\030\003 \001(\010\".\n\rMultiR" + - "esponse\022\035\n\006result\030\001 \003(\0132\r.ActionResult2\342" + - "\002\n\rClientService\022 \n\003get\022\013.GetRequest\032\014.G" + - "etResponse\022/\n\010multiGet\022\020.MultiGetRequest" + - "\032\021.MultiGetResponse\022)\n\006mutate\022\016.MutateRe" + - "quest\032\017.MutateResponse\022#\n\004scan\022\014.ScanReq" + - "uest\032\r.ScanResponse\022>\n\rbulkLoadHFile\022\025.B" + - "ulkLoadHFileRequest\032\026.BulkLoadHFileRespo" + - "nse\022F\n\013execService\022\032.CoprocessorServiceR" + - "equest\032\033.CoprocessorServiceResponse\022&\n\005m", - "ulti\022\r.MultiRequest\032\016.MultiResponseBB\n*o" + - "rg.apache.hadoop.hbase.protobuf.generate" + - "dB\014ClientProtosH\001\210\001\001\240\001\001" + "\n\014Client.proto\032\013hbase.proto\032\014Filter.prot" + + "o\032\nCell.proto\032\020Comparator.proto\"+\n\006Colum" + + "n\022\016\n\006family\030\001 \002(\014\022\021\n\tqualifier\030\002 \003(\014\"\342\001\n" + + "\003Get\022\013\n\003row\030\001 \002(\014\022\027\n\006column\030\002 \003(\0132\007.Colu" + + "mn\022!\n\tattribute\030\003 \003(\0132\016.NameBytesPair\022\027\n" + + "\006filter\030\004 \001(\0132\007.Filter\022\035\n\ttimeRange\030\005 \001(" + + "\0132\n.TimeRange\022\026\n\013maxVersions\030\006 \001(\r:\0011\022\031\n" + + "\013cacheBlocks\030\007 \001(\010:\004true\022\022\n\nstoreLimit\030\010" + + " \001(\r\022\023\n\013storeOffset\030\t \001(\r\":\n\006Result\022\023\n\004c" + + "ell\030\001 \003(\0132\005.Cell\022\033\n\023associatedCellCount\030", + "\002 \001(\005\"r\n\nGetRequest\022 \n\006region\030\001 \002(\0132\020.Re" + + "gionSpecifier\022\021\n\003get\030\002 \002(\0132\004.Get\022\030\n\020clos" + + "estRowBefore\030\003 \001(\010\022\025\n\rexistenceOnly\030\004 \001(" + + "\010\"w\n\017MultiGetRequest\022 \n\006region\030\001 \002(\0132\020.R" + + "egionSpecifier\022\021\n\003get\030\002 \003(\0132\004.Get\022\030\n\020clo" + + "sestRowBefore\030\003 \001(\010\022\025\n\rexistenceOnly\030\004 \001" + + "(\010\"6\n\013GetResponse\022\027\n\006result\030\001 \001(\0132\007.Resu" + + "lt\022\016\n\006exists\030\002 \001(\010\";\n\020MultiGetResponse\022\027" + + "\n\006result\030\001 \003(\0132\007.Result\022\016\n\006exists\030\002 \003(\010\"" + + "\177\n\tCondition\022\013\n\003row\030\001 \002(\014\022\016\n\006family\030\002 \002(", + "\014\022\021\n\tqualifier\030\003 \002(\014\022!\n\013compareType\030\004 \002(" + + "\0162\014.CompareType\022\037\n\ncomparator\030\005 \002(\0132\013.Co" + + "mparator\"\365\005\n\rMutationProto\022\013\n\003row\030\001 \001(\014\022" + + "/\n\nmutateType\030\002 \001(\0162\033.MutationProto.Muta" + + "tionType\022/\n\013columnValue\030\003 \003(\0132\032.Mutation" + + "Proto.ColumnValue\022\021\n\ttimestamp\030\004 \001(\004\022!\n\t" + + "attribute\030\005 \003(\0132\016.NameBytesPair\022:\n\ndurab" + + "ility\030\006 \001(\0162\031.MutationProto.Durability:\013" + + "USE_DEFAULT\022\035\n\ttimeRange\030\007 \001(\0132\n.TimeRan" + + "ge\022\033\n\023associatedCellCount\030\010 \001(\005\032\326\001\n\013Colu", + "mnValue\022\016\n\006family\030\001 \002(\014\022A\n\016qualifierValu" + + "e\030\002 \003(\0132).MutationProto.ColumnValue.Qual" + + "ifierValue\032t\n\016QualifierValue\022\021\n\tqualifie" + + "r\030\001 \001(\014\022\r\n\005value\030\002 \001(\014\022\021\n\ttimestamp\030\003 \001(" + + "\004\022-\n\ndeleteType\030\004 \001(\0162\031.MutationProto.De" + + "leteType\"W\n\nDurability\022\017\n\013USE_DEFAULT\020\000\022" + + "\014\n\010SKIP_WAL\020\001\022\r\n\tASYNC_WAL\020\002\022\014\n\010SYNC_WAL" + + "\020\003\022\r\n\tFSYNC_WAL\020\004\">\n\014MutationType\022\n\n\006APP" + + "END\020\000\022\r\n\tINCREMENT\020\001\022\007\n\003PUT\020\002\022\n\n\006DELETE\020" + + "\003\"U\n\nDeleteType\022\026\n\022DELETE_ONE_VERSION\020\000\022", + "\034\n\030DELETE_MULTIPLE_VERSIONS\020\001\022\021\n\rDELETE_" + + "FAMILY\020\002\"r\n\rMutateRequest\022 \n\006region\030\001 \002(" + + "\0132\020.RegionSpecifier\022 \n\010mutation\030\002 \002(\0132\016." + + "MutationProto\022\035\n\tcondition\030\003 \001(\0132\n.Condi" + + "tion\"<\n\016MutateResponse\022\027\n\006result\030\001 \001(\0132\007" + + ".Result\022\021\n\tprocessed\030\002 \001(\010\"\362\002\n\004Scan\022\027\n\006c" + + "olumn\030\001 \003(\0132\007.Column\022!\n\tattribute\030\002 \003(\0132" + + "\016.NameBytesPair\022\020\n\010startRow\030\003 \001(\014\022\017\n\007sto" + + "pRow\030\004 \001(\014\022\027\n\006filter\030\005 \001(\0132\007.Filter\022\035\n\tt" + + "imeRange\030\006 \001(\0132\n.TimeRange\022\026\n\013maxVersion", + "s\030\007 \001(\r:\0011\022\031\n\013cacheBlocks\030\010 \001(\010:\004true\022\021\n" + + "\tbatchSize\030\t \001(\r\022\025\n\rmaxResultSize\030\n \001(\004\022" + + "\022\n\nstoreLimit\030\013 \001(\r\022\023\n\013storeOffset\030\014 \001(\r" + + "\022\"\n\032loadColumnFamiliesOnDemand\030\r \001(\010\022\024\n\014" + + "cachingCount\030\016 \001(\r\022\023\n\013prefetching\030\017 \001(\010\"" + + "\230\001\n\013ScanRequest\022 \n\006region\030\001 \001(\0132\020.Region" + + "Specifier\022\023\n\004scan\030\002 \001(\0132\005.Scan\022\021\n\tscanne" + + "rId\030\003 \001(\004\022\024\n\014numberOfRows\030\004 \001(\r\022\024\n\014close" + + "Scanner\030\005 \001(\010\022\023\n\013nextCallSeq\030\006 \001(\004\"l\n\014Sc" + + "anResponse\022\'\n\016resultCellMeta\030\001 \001(\0132\017.Res", + "ultCellMeta\022\021\n\tscannerId\030\002 \001(\004\022\023\n\013moreRe" + + "sults\030\003 \001(\010\022\013\n\003ttl\030\004 \001(\r\"%\n\016ResultCellMe" + + "ta\022\023\n\013cellsLength\030\001 \003(\r\"\260\001\n\024BulkLoadHFil" + + "eRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpecifi" + + "er\0224\n\nfamilyPath\030\002 \003(\0132 .BulkLoadHFileRe" + + "quest.FamilyPath\022\024\n\014assignSeqNum\030\003 \001(\010\032*" + + "\n\nFamilyPath\022\016\n\006family\030\001 \002(\014\022\014\n\004path\030\002 \002" + + "(\t\"\'\n\025BulkLoadHFileResponse\022\016\n\006loaded\030\001 " + + "\002(\010\"_\n\026CoprocessorServiceCall\022\013\n\003row\030\001 \002" + + "(\014\022\023\n\013serviceName\030\002 \002(\t\022\022\n\nmethodName\030\003 ", + "\002(\t\022\017\n\007request\030\004 \002(\014\"d\n\031CoprocessorServi" + + "ceRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpecif" + + "ier\022%\n\004call\030\002 \002(\0132\027.CoprocessorServiceCa" + + "ll\"]\n\032CoprocessorServiceResponse\022 \n\006regi" + + "on\030\001 \002(\0132\020.RegionSpecifier\022\035\n\005value\030\002 \002(" + + "\0132\016.NameBytesPair\"B\n\013MultiAction\022 \n\010muta" + + "tion\030\001 \001(\0132\016.MutationProto\022\021\n\003get\030\002 \001(\0132" + + "\004.Get\"I\n\014ActionResult\022\026\n\005value\030\001 \001(\0132\007.R" + + "esult\022!\n\texception\030\002 \001(\0132\016.NameBytesPair" + + "\"^\n\014MultiRequest\022 \n\006region\030\001 \002(\0132\020.Regio", + "nSpecifier\022\034\n\006action\030\002 \003(\0132\014.MultiAction" + + "\022\016\n\006atomic\030\003 \001(\010\".\n\rMultiResponse\022\035\n\006res" + + "ult\030\001 \003(\0132\r.ActionResult2\342\002\n\rClientServi" + + "ce\022 \n\003get\022\013.GetRequest\032\014.GetResponse\022/\n\010" + + "multiGet\022\020.MultiGetRequest\032\021.MultiGetRes" + + "ponse\022)\n\006mutate\022\016.MutateRequest\032\017.Mutate" + + "Response\022#\n\004scan\022\014.ScanRequest\032\r.ScanRes" + + "ponse\022>\n\rbulkLoadHFile\022\025.BulkLoadHFileRe" + + "quest\032\026.BulkLoadHFileResponse\022F\n\013execSer" + + "vice\022\032.CoprocessorServiceRequest\032\033.Copro", + "cessorServiceResponse\022&\n\005multi\022\r.MultiRe" + + "quest\032\016.MultiResponseBB\n*org.apache.hado" + + "op.hbase.protobuf.generatedB\014ClientProto" + + "sH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -21876,6 +21877,8 @@ public final class ClientProtos { .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.getDescriptor(), + org.apache.hadoop.hbase.protobuf.generated.CellProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.getDescriptor(), }, assigner); } diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java index 6dd12e3..196d00c 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java @@ -1367,6 +1367,2752 @@ public final class ClusterStatusProtos { // @@protoc_insertion_point(class_scope:RegionInTransition) } + public interface RegionLoadOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // required .RegionSpecifier regionSpecifier = 1; + boolean hasRegionSpecifier(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder(); + + // optional uint32 stores = 2; + boolean hasStores(); + int getStores(); + + // optional uint32 storefiles = 3; + boolean hasStorefiles(); + int getStorefiles(); + + // optional uint32 storeUncompressedSizeMB = 4; + boolean hasStoreUncompressedSizeMB(); + int getStoreUncompressedSizeMB(); + + // optional uint32 storefileSizeMB = 5; + boolean hasStorefileSizeMB(); + int getStorefileSizeMB(); + + // optional uint32 memstoreSizeMB = 6; + boolean hasMemstoreSizeMB(); + int getMemstoreSizeMB(); + + // optional uint32 storefileIndexSizeMB = 7; + boolean hasStorefileIndexSizeMB(); + int getStorefileIndexSizeMB(); + + // optional uint64 readRequestsCount = 8; + boolean hasReadRequestsCount(); + long getReadRequestsCount(); + + // optional uint64 writeRequestsCount = 9; + boolean hasWriteRequestsCount(); + long getWriteRequestsCount(); + + // optional uint64 totalCompactingKVs = 10; + boolean hasTotalCompactingKVs(); + long getTotalCompactingKVs(); + + // optional uint64 currentCompactedKVs = 11; + boolean hasCurrentCompactedKVs(); + long getCurrentCompactedKVs(); + + // optional uint32 rootIndexSizeKB = 12; + boolean hasRootIndexSizeKB(); + int getRootIndexSizeKB(); + + // optional uint32 totalStaticIndexSizeKB = 13; + boolean hasTotalStaticIndexSizeKB(); + int getTotalStaticIndexSizeKB(); + + // optional uint32 totalStaticBloomSizeKB = 14; + boolean hasTotalStaticBloomSizeKB(); + int getTotalStaticBloomSizeKB(); + + // optional uint64 completeSequenceId = 15; + boolean hasCompleteSequenceId(); + long getCompleteSequenceId(); + } + public static final class RegionLoad extends + com.google.protobuf.GeneratedMessage + implements RegionLoadOrBuilder { + // Use RegionLoad.newBuilder() to construct. + private RegionLoad(Builder builder) { + super(builder); + } + private RegionLoad(boolean noInit) {} + + private static final RegionLoad defaultInstance; + public static RegionLoad getDefaultInstance() { + return defaultInstance; + } + + public RegionLoad getDefaultInstanceForType() { + return defaultInstance; + } + + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_RegionLoad_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_RegionLoad_fieldAccessorTable; + } + + private int bitField0_; + // required .RegionSpecifier regionSpecifier = 1; + public static final int REGIONSPECIFIER_FIELD_NUMBER = 1; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_; + public boolean hasRegionSpecifier() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier() { + return regionSpecifier_; + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder() { + return regionSpecifier_; + } + + // optional uint32 stores = 2; + public static final int STORES_FIELD_NUMBER = 2; + private int stores_; + public boolean hasStores() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public int getStores() { + return stores_; + } + + // optional uint32 storefiles = 3; + public static final int STOREFILES_FIELD_NUMBER = 3; + private int storefiles_; + public boolean hasStorefiles() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public int getStorefiles() { + return storefiles_; + } + + // optional uint32 storeUncompressedSizeMB = 4; + public static final int STOREUNCOMPRESSEDSIZEMB_FIELD_NUMBER = 4; + private int storeUncompressedSizeMB_; + public boolean hasStoreUncompressedSizeMB() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public int getStoreUncompressedSizeMB() { + return storeUncompressedSizeMB_; + } + + // optional uint32 storefileSizeMB = 5; + public static final int STOREFILESIZEMB_FIELD_NUMBER = 5; + private int storefileSizeMB_; + public boolean hasStorefileSizeMB() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public int getStorefileSizeMB() { + return storefileSizeMB_; + } + + // optional uint32 memstoreSizeMB = 6; + public static final int MEMSTORESIZEMB_FIELD_NUMBER = 6; + private int memstoreSizeMB_; + public boolean hasMemstoreSizeMB() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public int getMemstoreSizeMB() { + return memstoreSizeMB_; + } + + // optional uint32 storefileIndexSizeMB = 7; + public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 7; + private int storefileIndexSizeMB_; + public boolean hasStorefileIndexSizeMB() { + return ((bitField0_ & 0x00000040) == 0x00000040); + } + public int getStorefileIndexSizeMB() { + return storefileIndexSizeMB_; + } + + // optional uint64 readRequestsCount = 8; + public static final int READREQUESTSCOUNT_FIELD_NUMBER = 8; + private long readRequestsCount_; + public boolean hasReadRequestsCount() { + return ((bitField0_ & 0x00000080) == 0x00000080); + } + public long getReadRequestsCount() { + return readRequestsCount_; + } + + // optional uint64 writeRequestsCount = 9; + public static final int WRITEREQUESTSCOUNT_FIELD_NUMBER = 9; + private long writeRequestsCount_; + public boolean hasWriteRequestsCount() { + return ((bitField0_ & 0x00000100) == 0x00000100); + } + public long getWriteRequestsCount() { + return writeRequestsCount_; + } + + // optional uint64 totalCompactingKVs = 10; + public static final int TOTALCOMPACTINGKVS_FIELD_NUMBER = 10; + private long totalCompactingKVs_; + public boolean hasTotalCompactingKVs() { + return ((bitField0_ & 0x00000200) == 0x00000200); + } + public long getTotalCompactingKVs() { + return totalCompactingKVs_; + } + + // optional uint64 currentCompactedKVs = 11; + public static final int CURRENTCOMPACTEDKVS_FIELD_NUMBER = 11; + private long currentCompactedKVs_; + public boolean hasCurrentCompactedKVs() { + return ((bitField0_ & 0x00000400) == 0x00000400); + } + public long getCurrentCompactedKVs() { + return currentCompactedKVs_; + } + + // optional uint32 rootIndexSizeKB = 12; + public static final int ROOTINDEXSIZEKB_FIELD_NUMBER = 12; + private int rootIndexSizeKB_; + public boolean hasRootIndexSizeKB() { + return ((bitField0_ & 0x00000800) == 0x00000800); + } + public int getRootIndexSizeKB() { + return rootIndexSizeKB_; + } + + // optional uint32 totalStaticIndexSizeKB = 13; + public static final int TOTALSTATICINDEXSIZEKB_FIELD_NUMBER = 13; + private int totalStaticIndexSizeKB_; + public boolean hasTotalStaticIndexSizeKB() { + return ((bitField0_ & 0x00001000) == 0x00001000); + } + public int getTotalStaticIndexSizeKB() { + return totalStaticIndexSizeKB_; + } + + // optional uint32 totalStaticBloomSizeKB = 14; + public static final int TOTALSTATICBLOOMSIZEKB_FIELD_NUMBER = 14; + private int totalStaticBloomSizeKB_; + public boolean hasTotalStaticBloomSizeKB() { + return ((bitField0_ & 0x00002000) == 0x00002000); + } + public int getTotalStaticBloomSizeKB() { + return totalStaticBloomSizeKB_; + } + + // optional uint64 completeSequenceId = 15; + public static final int COMPLETESEQUENCEID_FIELD_NUMBER = 15; + private long completeSequenceId_; + public boolean hasCompleteSequenceId() { + return ((bitField0_ & 0x00004000) == 0x00004000); + } + public long getCompleteSequenceId() { + return completeSequenceId_; + } + + private void initFields() { + regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + stores_ = 0; + storefiles_ = 0; + storeUncompressedSizeMB_ = 0; + storefileSizeMB_ = 0; + memstoreSizeMB_ = 0; + storefileIndexSizeMB_ = 0; + readRequestsCount_ = 0L; + writeRequestsCount_ = 0L; + totalCompactingKVs_ = 0L; + currentCompactedKVs_ = 0L; + rootIndexSizeKB_ = 0; + totalStaticIndexSizeKB_ = 0; + totalStaticBloomSizeKB_ = 0; + completeSequenceId_ = 0L; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + if (!hasRegionSpecifier()) { + memoizedIsInitialized = 0; + return false; + } + if (!getRegionSpecifier().isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeMessage(1, regionSpecifier_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeUInt32(2, stores_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeUInt32(3, storefiles_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeUInt32(4, storeUncompressedSizeMB_); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + output.writeUInt32(5, storefileSizeMB_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + output.writeUInt32(6, memstoreSizeMB_); + } + if (((bitField0_ & 0x00000040) == 0x00000040)) { + output.writeUInt32(7, storefileIndexSizeMB_); + } + if (((bitField0_ & 0x00000080) == 0x00000080)) { + output.writeUInt64(8, readRequestsCount_); + } + if (((bitField0_ & 0x00000100) == 0x00000100)) { + output.writeUInt64(9, writeRequestsCount_); + } + if (((bitField0_ & 0x00000200) == 0x00000200)) { + output.writeUInt64(10, totalCompactingKVs_); + } + if (((bitField0_ & 0x00000400) == 0x00000400)) { + output.writeUInt64(11, currentCompactedKVs_); + } + if (((bitField0_ & 0x00000800) == 0x00000800)) { + output.writeUInt32(12, rootIndexSizeKB_); + } + if (((bitField0_ & 0x00001000) == 0x00001000)) { + output.writeUInt32(13, totalStaticIndexSizeKB_); + } + if (((bitField0_ & 0x00002000) == 0x00002000)) { + output.writeUInt32(14, totalStaticBloomSizeKB_); + } + if (((bitField0_ & 0x00004000) == 0x00004000)) { + output.writeUInt64(15, completeSequenceId_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeMessageSize(1, regionSpecifier_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(2, stores_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(3, storefiles_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(4, storeUncompressedSizeMB_); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(5, storefileSizeMB_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(6, memstoreSizeMB_); + } + if (((bitField0_ & 0x00000040) == 0x00000040)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(7, storefileIndexSizeMB_); + } + if (((bitField0_ & 0x00000080) == 0x00000080)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(8, readRequestsCount_); + } + if (((bitField0_ & 0x00000100) == 0x00000100)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(9, writeRequestsCount_); + } + if (((bitField0_ & 0x00000200) == 0x00000200)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(10, totalCompactingKVs_); + } + if (((bitField0_ & 0x00000400) == 0x00000400)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(11, currentCompactedKVs_); + } + if (((bitField0_ & 0x00000800) == 0x00000800)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(12, rootIndexSizeKB_); + } + if (((bitField0_ & 0x00001000) == 0x00001000)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(13, totalStaticIndexSizeKB_); + } + if (((bitField0_ & 0x00002000) == 0x00002000)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(14, totalStaticBloomSizeKB_); + } + if (((bitField0_ & 0x00004000) == 0x00004000)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(15, completeSequenceId_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad other = (org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad) obj; + + boolean result = true; + result = result && (hasRegionSpecifier() == other.hasRegionSpecifier()); + if (hasRegionSpecifier()) { + result = result && getRegionSpecifier() + .equals(other.getRegionSpecifier()); + } + result = result && (hasStores() == other.hasStores()); + if (hasStores()) { + result = result && (getStores() + == other.getStores()); + } + result = result && (hasStorefiles() == other.hasStorefiles()); + if (hasStorefiles()) { + result = result && (getStorefiles() + == other.getStorefiles()); + } + result = result && (hasStoreUncompressedSizeMB() == other.hasStoreUncompressedSizeMB()); + if (hasStoreUncompressedSizeMB()) { + result = result && (getStoreUncompressedSizeMB() + == other.getStoreUncompressedSizeMB()); + } + result = result && (hasStorefileSizeMB() == other.hasStorefileSizeMB()); + if (hasStorefileSizeMB()) { + result = result && (getStorefileSizeMB() + == other.getStorefileSizeMB()); + } + result = result && (hasMemstoreSizeMB() == other.hasMemstoreSizeMB()); + if (hasMemstoreSizeMB()) { + result = result && (getMemstoreSizeMB() + == other.getMemstoreSizeMB()); + } + result = result && (hasStorefileIndexSizeMB() == other.hasStorefileIndexSizeMB()); + if (hasStorefileIndexSizeMB()) { + result = result && (getStorefileIndexSizeMB() + == other.getStorefileIndexSizeMB()); + } + result = result && (hasReadRequestsCount() == other.hasReadRequestsCount()); + if (hasReadRequestsCount()) { + result = result && (getReadRequestsCount() + == other.getReadRequestsCount()); + } + result = result && (hasWriteRequestsCount() == other.hasWriteRequestsCount()); + if (hasWriteRequestsCount()) { + result = result && (getWriteRequestsCount() + == other.getWriteRequestsCount()); + } + result = result && (hasTotalCompactingKVs() == other.hasTotalCompactingKVs()); + if (hasTotalCompactingKVs()) { + result = result && (getTotalCompactingKVs() + == other.getTotalCompactingKVs()); + } + result = result && (hasCurrentCompactedKVs() == other.hasCurrentCompactedKVs()); + if (hasCurrentCompactedKVs()) { + result = result && (getCurrentCompactedKVs() + == other.getCurrentCompactedKVs()); + } + result = result && (hasRootIndexSizeKB() == other.hasRootIndexSizeKB()); + if (hasRootIndexSizeKB()) { + result = result && (getRootIndexSizeKB() + == other.getRootIndexSizeKB()); + } + result = result && (hasTotalStaticIndexSizeKB() == other.hasTotalStaticIndexSizeKB()); + if (hasTotalStaticIndexSizeKB()) { + result = result && (getTotalStaticIndexSizeKB() + == other.getTotalStaticIndexSizeKB()); + } + result = result && (hasTotalStaticBloomSizeKB() == other.hasTotalStaticBloomSizeKB()); + if (hasTotalStaticBloomSizeKB()) { + result = result && (getTotalStaticBloomSizeKB() + == other.getTotalStaticBloomSizeKB()); + } + result = result && (hasCompleteSequenceId() == other.hasCompleteSequenceId()); + if (hasCompleteSequenceId()) { + result = result && (getCompleteSequenceId() + == other.getCompleteSequenceId()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + @java.lang.Override + public int hashCode() { + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasRegionSpecifier()) { + hash = (37 * hash) + REGIONSPECIFIER_FIELD_NUMBER; + hash = (53 * hash) + getRegionSpecifier().hashCode(); + } + if (hasStores()) { + hash = (37 * hash) + STORES_FIELD_NUMBER; + hash = (53 * hash) + getStores(); + } + if (hasStorefiles()) { + hash = (37 * hash) + STOREFILES_FIELD_NUMBER; + hash = (53 * hash) + getStorefiles(); + } + if (hasStoreUncompressedSizeMB()) { + hash = (37 * hash) + STOREUNCOMPRESSEDSIZEMB_FIELD_NUMBER; + hash = (53 * hash) + getStoreUncompressedSizeMB(); + } + if (hasStorefileSizeMB()) { + hash = (37 * hash) + STOREFILESIZEMB_FIELD_NUMBER; + hash = (53 * hash) + getStorefileSizeMB(); + } + if (hasMemstoreSizeMB()) { + hash = (37 * hash) + MEMSTORESIZEMB_FIELD_NUMBER; + hash = (53 * hash) + getMemstoreSizeMB(); + } + if (hasStorefileIndexSizeMB()) { + hash = (37 * hash) + STOREFILEINDEXSIZEMB_FIELD_NUMBER; + hash = (53 * hash) + getStorefileIndexSizeMB(); + } + if (hasReadRequestsCount()) { + hash = (37 * hash) + READREQUESTSCOUNT_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getReadRequestsCount()); + } + if (hasWriteRequestsCount()) { + hash = (37 * hash) + WRITEREQUESTSCOUNT_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getWriteRequestsCount()); + } + if (hasTotalCompactingKVs()) { + hash = (37 * hash) + TOTALCOMPACTINGKVS_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getTotalCompactingKVs()); + } + if (hasCurrentCompactedKVs()) { + hash = (37 * hash) + CURRENTCOMPACTEDKVS_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getCurrentCompactedKVs()); + } + if (hasRootIndexSizeKB()) { + hash = (37 * hash) + ROOTINDEXSIZEKB_FIELD_NUMBER; + hash = (53 * hash) + getRootIndexSizeKB(); + } + if (hasTotalStaticIndexSizeKB()) { + hash = (37 * hash) + TOTALSTATICINDEXSIZEKB_FIELD_NUMBER; + hash = (53 * hash) + getTotalStaticIndexSizeKB(); + } + if (hasTotalStaticBloomSizeKB()) { + hash = (37 * hash) + TOTALSTATICBLOOMSIZEKB_FIELD_NUMBER; + hash = (53 * hash) + getTotalStaticBloomSizeKB(); + } + if (hasCompleteSequenceId()) { + hash = (37 * hash) + COMPLETESEQUENCEID_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getCompleteSequenceId()); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom(java.io.InputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input, extensionRegistry)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_RegionLoad_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_RegionLoad_fieldAccessorTable; + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + getRegionSpecifierFieldBuilder(); + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + if (regionSpecifierBuilder_ == null) { + regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + } else { + regionSpecifierBuilder_.clear(); + } + bitField0_ = (bitField0_ & ~0x00000001); + stores_ = 0; + bitField0_ = (bitField0_ & ~0x00000002); + storefiles_ = 0; + bitField0_ = (bitField0_ & ~0x00000004); + storeUncompressedSizeMB_ = 0; + bitField0_ = (bitField0_ & ~0x00000008); + storefileSizeMB_ = 0; + bitField0_ = (bitField0_ & ~0x00000010); + memstoreSizeMB_ = 0; + bitField0_ = (bitField0_ & ~0x00000020); + storefileIndexSizeMB_ = 0; + bitField0_ = (bitField0_ & ~0x00000040); + readRequestsCount_ = 0L; + bitField0_ = (bitField0_ & ~0x00000080); + writeRequestsCount_ = 0L; + bitField0_ = (bitField0_ & ~0x00000100); + totalCompactingKVs_ = 0L; + bitField0_ = (bitField0_ & ~0x00000200); + currentCompactedKVs_ = 0L; + bitField0_ = (bitField0_ & ~0x00000400); + rootIndexSizeKB_ = 0; + bitField0_ = (bitField0_ & ~0x00000800); + totalStaticIndexSizeKB_ = 0; + bitField0_ = (bitField0_ & ~0x00001000); + totalStaticBloomSizeKB_ = 0; + bitField0_ = (bitField0_ & ~0x00002000); + completeSequenceId_ = 0L; + bitField0_ = (bitField0_ & ~0x00004000); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.getDescriptor(); + } + + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad build() { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad buildParsed() + throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException( + result).asInvalidProtocolBufferException(); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad result = new org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + if (regionSpecifierBuilder_ == null) { + result.regionSpecifier_ = regionSpecifier_; + } else { + result.regionSpecifier_ = regionSpecifierBuilder_.build(); + } + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.stores_ = stores_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.storefiles_ = storefiles_; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000008; + } + result.storeUncompressedSizeMB_ = storeUncompressedSizeMB_; + if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + to_bitField0_ |= 0x00000010; + } + result.storefileSizeMB_ = storefileSizeMB_; + if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + to_bitField0_ |= 0x00000020; + } + result.memstoreSizeMB_ = memstoreSizeMB_; + if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + to_bitField0_ |= 0x00000040; + } + result.storefileIndexSizeMB_ = storefileIndexSizeMB_; + if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + to_bitField0_ |= 0x00000080; + } + result.readRequestsCount_ = readRequestsCount_; + if (((from_bitField0_ & 0x00000100) == 0x00000100)) { + to_bitField0_ |= 0x00000100; + } + result.writeRequestsCount_ = writeRequestsCount_; + if (((from_bitField0_ & 0x00000200) == 0x00000200)) { + to_bitField0_ |= 0x00000200; + } + result.totalCompactingKVs_ = totalCompactingKVs_; + if (((from_bitField0_ & 0x00000400) == 0x00000400)) { + to_bitField0_ |= 0x00000400; + } + result.currentCompactedKVs_ = currentCompactedKVs_; + if (((from_bitField0_ & 0x00000800) == 0x00000800)) { + to_bitField0_ |= 0x00000800; + } + result.rootIndexSizeKB_ = rootIndexSizeKB_; + if (((from_bitField0_ & 0x00001000) == 0x00001000)) { + to_bitField0_ |= 0x00001000; + } + result.totalStaticIndexSizeKB_ = totalStaticIndexSizeKB_; + if (((from_bitField0_ & 0x00002000) == 0x00002000)) { + to_bitField0_ |= 0x00002000; + } + result.totalStaticBloomSizeKB_ = totalStaticBloomSizeKB_; + if (((from_bitField0_ & 0x00004000) == 0x00004000)) { + to_bitField0_ |= 0x00004000; + } + result.completeSequenceId_ = completeSequenceId_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.getDefaultInstance()) return this; + if (other.hasRegionSpecifier()) { + mergeRegionSpecifier(other.getRegionSpecifier()); + } + if (other.hasStores()) { + setStores(other.getStores()); + } + if (other.hasStorefiles()) { + setStorefiles(other.getStorefiles()); + } + if (other.hasStoreUncompressedSizeMB()) { + setStoreUncompressedSizeMB(other.getStoreUncompressedSizeMB()); + } + if (other.hasStorefileSizeMB()) { + setStorefileSizeMB(other.getStorefileSizeMB()); + } + if (other.hasMemstoreSizeMB()) { + setMemstoreSizeMB(other.getMemstoreSizeMB()); + } + if (other.hasStorefileIndexSizeMB()) { + setStorefileIndexSizeMB(other.getStorefileIndexSizeMB()); + } + if (other.hasReadRequestsCount()) { + setReadRequestsCount(other.getReadRequestsCount()); + } + if (other.hasWriteRequestsCount()) { + setWriteRequestsCount(other.getWriteRequestsCount()); + } + if (other.hasTotalCompactingKVs()) { + setTotalCompactingKVs(other.getTotalCompactingKVs()); + } + if (other.hasCurrentCompactedKVs()) { + setCurrentCompactedKVs(other.getCurrentCompactedKVs()); + } + if (other.hasRootIndexSizeKB()) { + setRootIndexSizeKB(other.getRootIndexSizeKB()); + } + if (other.hasTotalStaticIndexSizeKB()) { + setTotalStaticIndexSizeKB(other.getTotalStaticIndexSizeKB()); + } + if (other.hasTotalStaticBloomSizeKB()) { + setTotalStaticBloomSizeKB(other.getTotalStaticBloomSizeKB()); + } + if (other.hasCompleteSequenceId()) { + setCompleteSequenceId(other.getCompleteSequenceId()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + if (!hasRegionSpecifier()) { + + return false; + } + if (!getRegionSpecifier().isInitialized()) { + + return false; + } + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder( + this.getUnknownFields()); + while (true) { + int tag = input.readTag(); + switch (tag) { + case 0: + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + } + break; + } + case 10: { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(); + if (hasRegionSpecifier()) { + subBuilder.mergeFrom(getRegionSpecifier()); + } + input.readMessage(subBuilder, extensionRegistry); + setRegionSpecifier(subBuilder.buildPartial()); + break; + } + case 16: { + bitField0_ |= 0x00000002; + stores_ = input.readUInt32(); + break; + } + case 24: { + bitField0_ |= 0x00000004; + storefiles_ = input.readUInt32(); + break; + } + case 32: { + bitField0_ |= 0x00000008; + storeUncompressedSizeMB_ = input.readUInt32(); + break; + } + case 40: { + bitField0_ |= 0x00000010; + storefileSizeMB_ = input.readUInt32(); + break; + } + case 48: { + bitField0_ |= 0x00000020; + memstoreSizeMB_ = input.readUInt32(); + break; + } + case 56: { + bitField0_ |= 0x00000040; + storefileIndexSizeMB_ = input.readUInt32(); + break; + } + case 64: { + bitField0_ |= 0x00000080; + readRequestsCount_ = input.readUInt64(); + break; + } + case 72: { + bitField0_ |= 0x00000100; + writeRequestsCount_ = input.readUInt64(); + break; + } + case 80: { + bitField0_ |= 0x00000200; + totalCompactingKVs_ = input.readUInt64(); + break; + } + case 88: { + bitField0_ |= 0x00000400; + currentCompactedKVs_ = input.readUInt64(); + break; + } + case 96: { + bitField0_ |= 0x00000800; + rootIndexSizeKB_ = input.readUInt32(); + break; + } + case 104: { + bitField0_ |= 0x00001000; + totalStaticIndexSizeKB_ = input.readUInt32(); + break; + } + case 112: { + bitField0_ |= 0x00002000; + totalStaticBloomSizeKB_ = input.readUInt32(); + break; + } + case 120: { + bitField0_ |= 0x00004000; + completeSequenceId_ = input.readUInt64(); + break; + } + } + } + } + + private int bitField0_; + + // required .RegionSpecifier regionSpecifier = 1; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + private com.google.protobuf.SingleFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionSpecifierBuilder_; + public boolean hasRegionSpecifier() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier() { + if (regionSpecifierBuilder_ == null) { + return regionSpecifier_; + } else { + return regionSpecifierBuilder_.getMessage(); + } + } + public Builder setRegionSpecifier(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { + if (regionSpecifierBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + regionSpecifier_ = value; + onChanged(); + } else { + regionSpecifierBuilder_.setMessage(value); + } + bitField0_ |= 0x00000001; + return this; + } + public Builder setRegionSpecifier( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { + if (regionSpecifierBuilder_ == null) { + regionSpecifier_ = builderForValue.build(); + onChanged(); + } else { + regionSpecifierBuilder_.setMessage(builderForValue.build()); + } + bitField0_ |= 0x00000001; + return this; + } + public Builder mergeRegionSpecifier(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { + if (regionSpecifierBuilder_ == null) { + if (((bitField0_ & 0x00000001) == 0x00000001) && + regionSpecifier_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { + regionSpecifier_ = + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionSpecifier_).mergeFrom(value).buildPartial(); + } else { + regionSpecifier_ = value; + } + onChanged(); + } else { + regionSpecifierBuilder_.mergeFrom(value); + } + bitField0_ |= 0x00000001; + return this; + } + public Builder clearRegionSpecifier() { + if (regionSpecifierBuilder_ == null) { + regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + onChanged(); + } else { + regionSpecifierBuilder_.clear(); + } + bitField0_ = (bitField0_ & ~0x00000001); + return this; + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionSpecifierBuilder() { + bitField0_ |= 0x00000001; + onChanged(); + return getRegionSpecifierFieldBuilder().getBuilder(); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder() { + if (regionSpecifierBuilder_ != null) { + return regionSpecifierBuilder_.getMessageOrBuilder(); + } else { + return regionSpecifier_; + } + } + private com.google.protobuf.SingleFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> + getRegionSpecifierFieldBuilder() { + if (regionSpecifierBuilder_ == null) { + regionSpecifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( + regionSpecifier_, + getParentForChildren(), + isClean()); + regionSpecifier_ = null; + } + return regionSpecifierBuilder_; + } + + // optional uint32 stores = 2; + private int stores_ ; + public boolean hasStores() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public int getStores() { + return stores_; + } + public Builder setStores(int value) { + bitField0_ |= 0x00000002; + stores_ = value; + onChanged(); + return this; + } + public Builder clearStores() { + bitField0_ = (bitField0_ & ~0x00000002); + stores_ = 0; + onChanged(); + return this; + } + + // optional uint32 storefiles = 3; + private int storefiles_ ; + public boolean hasStorefiles() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public int getStorefiles() { + return storefiles_; + } + public Builder setStorefiles(int value) { + bitField0_ |= 0x00000004; + storefiles_ = value; + onChanged(); + return this; + } + public Builder clearStorefiles() { + bitField0_ = (bitField0_ & ~0x00000004); + storefiles_ = 0; + onChanged(); + return this; + } + + // optional uint32 storeUncompressedSizeMB = 4; + private int storeUncompressedSizeMB_ ; + public boolean hasStoreUncompressedSizeMB() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public int getStoreUncompressedSizeMB() { + return storeUncompressedSizeMB_; + } + public Builder setStoreUncompressedSizeMB(int value) { + bitField0_ |= 0x00000008; + storeUncompressedSizeMB_ = value; + onChanged(); + return this; + } + public Builder clearStoreUncompressedSizeMB() { + bitField0_ = (bitField0_ & ~0x00000008); + storeUncompressedSizeMB_ = 0; + onChanged(); + return this; + } + + // optional uint32 storefileSizeMB = 5; + private int storefileSizeMB_ ; + public boolean hasStorefileSizeMB() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public int getStorefileSizeMB() { + return storefileSizeMB_; + } + public Builder setStorefileSizeMB(int value) { + bitField0_ |= 0x00000010; + storefileSizeMB_ = value; + onChanged(); + return this; + } + public Builder clearStorefileSizeMB() { + bitField0_ = (bitField0_ & ~0x00000010); + storefileSizeMB_ = 0; + onChanged(); + return this; + } + + // optional uint32 memstoreSizeMB = 6; + private int memstoreSizeMB_ ; + public boolean hasMemstoreSizeMB() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public int getMemstoreSizeMB() { + return memstoreSizeMB_; + } + public Builder setMemstoreSizeMB(int value) { + bitField0_ |= 0x00000020; + memstoreSizeMB_ = value; + onChanged(); + return this; + } + public Builder clearMemstoreSizeMB() { + bitField0_ = (bitField0_ & ~0x00000020); + memstoreSizeMB_ = 0; + onChanged(); + return this; + } + + // optional uint32 storefileIndexSizeMB = 7; + private int storefileIndexSizeMB_ ; + public boolean hasStorefileIndexSizeMB() { + return ((bitField0_ & 0x00000040) == 0x00000040); + } + public int getStorefileIndexSizeMB() { + return storefileIndexSizeMB_; + } + public Builder setStorefileIndexSizeMB(int value) { + bitField0_ |= 0x00000040; + storefileIndexSizeMB_ = value; + onChanged(); + return this; + } + public Builder clearStorefileIndexSizeMB() { + bitField0_ = (bitField0_ & ~0x00000040); + storefileIndexSizeMB_ = 0; + onChanged(); + return this; + } + + // optional uint64 readRequestsCount = 8; + private long readRequestsCount_ ; + public boolean hasReadRequestsCount() { + return ((bitField0_ & 0x00000080) == 0x00000080); + } + public long getReadRequestsCount() { + return readRequestsCount_; + } + public Builder setReadRequestsCount(long value) { + bitField0_ |= 0x00000080; + readRequestsCount_ = value; + onChanged(); + return this; + } + public Builder clearReadRequestsCount() { + bitField0_ = (bitField0_ & ~0x00000080); + readRequestsCount_ = 0L; + onChanged(); + return this; + } + + // optional uint64 writeRequestsCount = 9; + private long writeRequestsCount_ ; + public boolean hasWriteRequestsCount() { + return ((bitField0_ & 0x00000100) == 0x00000100); + } + public long getWriteRequestsCount() { + return writeRequestsCount_; + } + public Builder setWriteRequestsCount(long value) { + bitField0_ |= 0x00000100; + writeRequestsCount_ = value; + onChanged(); + return this; + } + public Builder clearWriteRequestsCount() { + bitField0_ = (bitField0_ & ~0x00000100); + writeRequestsCount_ = 0L; + onChanged(); + return this; + } + + // optional uint64 totalCompactingKVs = 10; + private long totalCompactingKVs_ ; + public boolean hasTotalCompactingKVs() { + return ((bitField0_ & 0x00000200) == 0x00000200); + } + public long getTotalCompactingKVs() { + return totalCompactingKVs_; + } + public Builder setTotalCompactingKVs(long value) { + bitField0_ |= 0x00000200; + totalCompactingKVs_ = value; + onChanged(); + return this; + } + public Builder clearTotalCompactingKVs() { + bitField0_ = (bitField0_ & ~0x00000200); + totalCompactingKVs_ = 0L; + onChanged(); + return this; + } + + // optional uint64 currentCompactedKVs = 11; + private long currentCompactedKVs_ ; + public boolean hasCurrentCompactedKVs() { + return ((bitField0_ & 0x00000400) == 0x00000400); + } + public long getCurrentCompactedKVs() { + return currentCompactedKVs_; + } + public Builder setCurrentCompactedKVs(long value) { + bitField0_ |= 0x00000400; + currentCompactedKVs_ = value; + onChanged(); + return this; + } + public Builder clearCurrentCompactedKVs() { + bitField0_ = (bitField0_ & ~0x00000400); + currentCompactedKVs_ = 0L; + onChanged(); + return this; + } + + // optional uint32 rootIndexSizeKB = 12; + private int rootIndexSizeKB_ ; + public boolean hasRootIndexSizeKB() { + return ((bitField0_ & 0x00000800) == 0x00000800); + } + public int getRootIndexSizeKB() { + return rootIndexSizeKB_; + } + public Builder setRootIndexSizeKB(int value) { + bitField0_ |= 0x00000800; + rootIndexSizeKB_ = value; + onChanged(); + return this; + } + public Builder clearRootIndexSizeKB() { + bitField0_ = (bitField0_ & ~0x00000800); + rootIndexSizeKB_ = 0; + onChanged(); + return this; + } + + // optional uint32 totalStaticIndexSizeKB = 13; + private int totalStaticIndexSizeKB_ ; + public boolean hasTotalStaticIndexSizeKB() { + return ((bitField0_ & 0x00001000) == 0x00001000); + } + public int getTotalStaticIndexSizeKB() { + return totalStaticIndexSizeKB_; + } + public Builder setTotalStaticIndexSizeKB(int value) { + bitField0_ |= 0x00001000; + totalStaticIndexSizeKB_ = value; + onChanged(); + return this; + } + public Builder clearTotalStaticIndexSizeKB() { + bitField0_ = (bitField0_ & ~0x00001000); + totalStaticIndexSizeKB_ = 0; + onChanged(); + return this; + } + + // optional uint32 totalStaticBloomSizeKB = 14; + private int totalStaticBloomSizeKB_ ; + public boolean hasTotalStaticBloomSizeKB() { + return ((bitField0_ & 0x00002000) == 0x00002000); + } + public int getTotalStaticBloomSizeKB() { + return totalStaticBloomSizeKB_; + } + public Builder setTotalStaticBloomSizeKB(int value) { + bitField0_ |= 0x00002000; + totalStaticBloomSizeKB_ = value; + onChanged(); + return this; + } + public Builder clearTotalStaticBloomSizeKB() { + bitField0_ = (bitField0_ & ~0x00002000); + totalStaticBloomSizeKB_ = 0; + onChanged(); + return this; + } + + // optional uint64 completeSequenceId = 15; + private long completeSequenceId_ ; + public boolean hasCompleteSequenceId() { + return ((bitField0_ & 0x00004000) == 0x00004000); + } + public long getCompleteSequenceId() { + return completeSequenceId_; + } + public Builder setCompleteSequenceId(long value) { + bitField0_ |= 0x00004000; + completeSequenceId_ = value; + onChanged(); + return this; + } + public Builder clearCompleteSequenceId() { + bitField0_ = (bitField0_ & ~0x00004000); + completeSequenceId_ = 0L; + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:RegionLoad) + } + + static { + defaultInstance = new RegionLoad(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:RegionLoad) + } + + public interface ServerLoadOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // optional uint32 numberOfRequests = 1; + boolean hasNumberOfRequests(); + int getNumberOfRequests(); + + // optional uint32 totalNumberOfRequests = 2; + boolean hasTotalNumberOfRequests(); + int getTotalNumberOfRequests(); + + // optional uint32 usedHeapMB = 3; + boolean hasUsedHeapMB(); + int getUsedHeapMB(); + + // optional uint32 maxHeapMB = 4; + boolean hasMaxHeapMB(); + int getMaxHeapMB(); + + // repeated .RegionLoad regionLoads = 5; + java.util.List + getRegionLoadsList(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad getRegionLoads(int index); + int getRegionLoadsCount(); + java.util.List + getRegionLoadsOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder getRegionLoadsOrBuilder( + int index); + + // repeated .Coprocessor coprocessors = 6; + java.util.List + getCoprocessorsList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getCoprocessors(int index); + int getCoprocessorsCount(); + java.util.List + getCoprocessorsOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( + int index); + + // optional uint64 reportStartTime = 7; + boolean hasReportStartTime(); + long getReportStartTime(); + + // optional uint64 reportEndTime = 8; + boolean hasReportEndTime(); + long getReportEndTime(); + + // optional uint32 infoServerPort = 9; + boolean hasInfoServerPort(); + int getInfoServerPort(); + } + public static final class ServerLoad extends + com.google.protobuf.GeneratedMessage + implements ServerLoadOrBuilder { + // Use ServerLoad.newBuilder() to construct. + private ServerLoad(Builder builder) { + super(builder); + } + private ServerLoad(boolean noInit) {} + + private static final ServerLoad defaultInstance; + public static ServerLoad getDefaultInstance() { + return defaultInstance; + } + + public ServerLoad getDefaultInstanceForType() { + return defaultInstance; + } + + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_ServerLoad_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_ServerLoad_fieldAccessorTable; + } + + private int bitField0_; + // optional uint32 numberOfRequests = 1; + public static final int NUMBEROFREQUESTS_FIELD_NUMBER = 1; + private int numberOfRequests_; + public boolean hasNumberOfRequests() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public int getNumberOfRequests() { + return numberOfRequests_; + } + + // optional uint32 totalNumberOfRequests = 2; + public static final int TOTALNUMBEROFREQUESTS_FIELD_NUMBER = 2; + private int totalNumberOfRequests_; + public boolean hasTotalNumberOfRequests() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public int getTotalNumberOfRequests() { + return totalNumberOfRequests_; + } + + // optional uint32 usedHeapMB = 3; + public static final int USEDHEAPMB_FIELD_NUMBER = 3; + private int usedHeapMB_; + public boolean hasUsedHeapMB() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public int getUsedHeapMB() { + return usedHeapMB_; + } + + // optional uint32 maxHeapMB = 4; + public static final int MAXHEAPMB_FIELD_NUMBER = 4; + private int maxHeapMB_; + public boolean hasMaxHeapMB() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public int getMaxHeapMB() { + return maxHeapMB_; + } + + // repeated .RegionLoad regionLoads = 5; + public static final int REGIONLOADS_FIELD_NUMBER = 5; + private java.util.List regionLoads_; + public java.util.List getRegionLoadsList() { + return regionLoads_; + } + public java.util.List + getRegionLoadsOrBuilderList() { + return regionLoads_; + } + public int getRegionLoadsCount() { + return regionLoads_.size(); + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad getRegionLoads(int index) { + return regionLoads_.get(index); + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder getRegionLoadsOrBuilder( + int index) { + return regionLoads_.get(index); + } + + // repeated .Coprocessor coprocessors = 6; + public static final int COPROCESSORS_FIELD_NUMBER = 6; + private java.util.List coprocessors_; + public java.util.List getCoprocessorsList() { + return coprocessors_; + } + public java.util.List + getCoprocessorsOrBuilderList() { + return coprocessors_; + } + public int getCoprocessorsCount() { + return coprocessors_.size(); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getCoprocessors(int index) { + return coprocessors_.get(index); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( + int index) { + return coprocessors_.get(index); + } + + // optional uint64 reportStartTime = 7; + public static final int REPORTSTARTTIME_FIELD_NUMBER = 7; + private long reportStartTime_; + public boolean hasReportStartTime() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + public long getReportStartTime() { + return reportStartTime_; + } + + // optional uint64 reportEndTime = 8; + public static final int REPORTENDTIME_FIELD_NUMBER = 8; + private long reportEndTime_; + public boolean hasReportEndTime() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + public long getReportEndTime() { + return reportEndTime_; + } + + // optional uint32 infoServerPort = 9; + public static final int INFOSERVERPORT_FIELD_NUMBER = 9; + private int infoServerPort_; + public boolean hasInfoServerPort() { + return ((bitField0_ & 0x00000040) == 0x00000040); + } + public int getInfoServerPort() { + return infoServerPort_; + } + + private void initFields() { + numberOfRequests_ = 0; + totalNumberOfRequests_ = 0; + usedHeapMB_ = 0; + maxHeapMB_ = 0; + regionLoads_ = java.util.Collections.emptyList(); + coprocessors_ = java.util.Collections.emptyList(); + reportStartTime_ = 0L; + reportEndTime_ = 0L; + infoServerPort_ = 0; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + for (int i = 0; i < getRegionLoadsCount(); i++) { + if (!getRegionLoads(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } + for (int i = 0; i < getCoprocessorsCount(); i++) { + if (!getCoprocessors(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeUInt32(1, numberOfRequests_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeUInt32(2, totalNumberOfRequests_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeUInt32(3, usedHeapMB_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeUInt32(4, maxHeapMB_); + } + for (int i = 0; i < regionLoads_.size(); i++) { + output.writeMessage(5, regionLoads_.get(i)); + } + for (int i = 0; i < coprocessors_.size(); i++) { + output.writeMessage(6, coprocessors_.get(i)); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + output.writeUInt64(7, reportStartTime_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + output.writeUInt64(8, reportEndTime_); + } + if (((bitField0_ & 0x00000040) == 0x00000040)) { + output.writeUInt32(9, infoServerPort_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(1, numberOfRequests_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(2, totalNumberOfRequests_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(3, usedHeapMB_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(4, maxHeapMB_); + } + for (int i = 0; i < regionLoads_.size(); i++) { + size += com.google.protobuf.CodedOutputStream + .computeMessageSize(5, regionLoads_.get(i)); + } + for (int i = 0; i < coprocessors_.size(); i++) { + size += com.google.protobuf.CodedOutputStream + .computeMessageSize(6, coprocessors_.get(i)); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(7, reportStartTime_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(8, reportEndTime_); + } + if (((bitField0_ & 0x00000040) == 0x00000040)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt32Size(9, infoServerPort_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad other = (org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad) obj; + + boolean result = true; + result = result && (hasNumberOfRequests() == other.hasNumberOfRequests()); + if (hasNumberOfRequests()) { + result = result && (getNumberOfRequests() + == other.getNumberOfRequests()); + } + result = result && (hasTotalNumberOfRequests() == other.hasTotalNumberOfRequests()); + if (hasTotalNumberOfRequests()) { + result = result && (getTotalNumberOfRequests() + == other.getTotalNumberOfRequests()); + } + result = result && (hasUsedHeapMB() == other.hasUsedHeapMB()); + if (hasUsedHeapMB()) { + result = result && (getUsedHeapMB() + == other.getUsedHeapMB()); + } + result = result && (hasMaxHeapMB() == other.hasMaxHeapMB()); + if (hasMaxHeapMB()) { + result = result && (getMaxHeapMB() + == other.getMaxHeapMB()); + } + result = result && getRegionLoadsList() + .equals(other.getRegionLoadsList()); + result = result && getCoprocessorsList() + .equals(other.getCoprocessorsList()); + result = result && (hasReportStartTime() == other.hasReportStartTime()); + if (hasReportStartTime()) { + result = result && (getReportStartTime() + == other.getReportStartTime()); + } + result = result && (hasReportEndTime() == other.hasReportEndTime()); + if (hasReportEndTime()) { + result = result && (getReportEndTime() + == other.getReportEndTime()); + } + result = result && (hasInfoServerPort() == other.hasInfoServerPort()); + if (hasInfoServerPort()) { + result = result && (getInfoServerPort() + == other.getInfoServerPort()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + @java.lang.Override + public int hashCode() { + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasNumberOfRequests()) { + hash = (37 * hash) + NUMBEROFREQUESTS_FIELD_NUMBER; + hash = (53 * hash) + getNumberOfRequests(); + } + if (hasTotalNumberOfRequests()) { + hash = (37 * hash) + TOTALNUMBEROFREQUESTS_FIELD_NUMBER; + hash = (53 * hash) + getTotalNumberOfRequests(); + } + if (hasUsedHeapMB()) { + hash = (37 * hash) + USEDHEAPMB_FIELD_NUMBER; + hash = (53 * hash) + getUsedHeapMB(); + } + if (hasMaxHeapMB()) { + hash = (37 * hash) + MAXHEAPMB_FIELD_NUMBER; + hash = (53 * hash) + getMaxHeapMB(); + } + if (getRegionLoadsCount() > 0) { + hash = (37 * hash) + REGIONLOADS_FIELD_NUMBER; + hash = (53 * hash) + getRegionLoadsList().hashCode(); + } + if (getCoprocessorsCount() > 0) { + hash = (37 * hash) + COPROCESSORS_FIELD_NUMBER; + hash = (53 * hash) + getCoprocessorsList().hashCode(); + } + if (hasReportStartTime()) { + hash = (37 * hash) + REPORTSTARTTIME_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getReportStartTime()); + } + if (hasReportEndTime()) { + hash = (37 * hash) + REPORTENDTIME_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getReportEndTime()); + } + if (hasInfoServerPort()) { + hash = (37 * hash) + INFOSERVERPORT_FIELD_NUMBER; + hash = (53 * hash) + getInfoServerPort(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom(java.io.InputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input, extensionRegistry)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_ServerLoad_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.internal_static_ServerLoad_fieldAccessorTable; + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + getRegionLoadsFieldBuilder(); + getCoprocessorsFieldBuilder(); + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + numberOfRequests_ = 0; + bitField0_ = (bitField0_ & ~0x00000001); + totalNumberOfRequests_ = 0; + bitField0_ = (bitField0_ & ~0x00000002); + usedHeapMB_ = 0; + bitField0_ = (bitField0_ & ~0x00000004); + maxHeapMB_ = 0; + bitField0_ = (bitField0_ & ~0x00000008); + if (regionLoadsBuilder_ == null) { + regionLoads_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000010); + } else { + regionLoadsBuilder_.clear(); + } + if (coprocessorsBuilder_ == null) { + coprocessors_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000020); + } else { + coprocessorsBuilder_.clear(); + } + reportStartTime_ = 0L; + bitField0_ = (bitField0_ & ~0x00000040); + reportEndTime_ = 0L; + bitField0_ = (bitField0_ & ~0x00000080); + infoServerPort_ = 0; + bitField0_ = (bitField0_ & ~0x00000100); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDescriptor(); + } + + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad build() { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad buildParsed() + throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException( + result).asInvalidProtocolBufferException(); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad result = new org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.numberOfRequests_ = numberOfRequests_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.totalNumberOfRequests_ = totalNumberOfRequests_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.usedHeapMB_ = usedHeapMB_; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000008; + } + result.maxHeapMB_ = maxHeapMB_; + if (regionLoadsBuilder_ == null) { + if (((bitField0_ & 0x00000010) == 0x00000010)) { + regionLoads_ = java.util.Collections.unmodifiableList(regionLoads_); + bitField0_ = (bitField0_ & ~0x00000010); + } + result.regionLoads_ = regionLoads_; + } else { + result.regionLoads_ = regionLoadsBuilder_.build(); + } + if (coprocessorsBuilder_ == null) { + if (((bitField0_ & 0x00000020) == 0x00000020)) { + coprocessors_ = java.util.Collections.unmodifiableList(coprocessors_); + bitField0_ = (bitField0_ & ~0x00000020); + } + result.coprocessors_ = coprocessors_; + } else { + result.coprocessors_ = coprocessorsBuilder_.build(); + } + if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + to_bitField0_ |= 0x00000010; + } + result.reportStartTime_ = reportStartTime_; + if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + to_bitField0_ |= 0x00000020; + } + result.reportEndTime_ = reportEndTime_; + if (((from_bitField0_ & 0x00000100) == 0x00000100)) { + to_bitField0_ |= 0x00000040; + } + result.infoServerPort_ = infoServerPort_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance()) return this; + if (other.hasNumberOfRequests()) { + setNumberOfRequests(other.getNumberOfRequests()); + } + if (other.hasTotalNumberOfRequests()) { + setTotalNumberOfRequests(other.getTotalNumberOfRequests()); + } + if (other.hasUsedHeapMB()) { + setUsedHeapMB(other.getUsedHeapMB()); + } + if (other.hasMaxHeapMB()) { + setMaxHeapMB(other.getMaxHeapMB()); + } + if (regionLoadsBuilder_ == null) { + if (!other.regionLoads_.isEmpty()) { + if (regionLoads_.isEmpty()) { + regionLoads_ = other.regionLoads_; + bitField0_ = (bitField0_ & ~0x00000010); + } else { + ensureRegionLoadsIsMutable(); + regionLoads_.addAll(other.regionLoads_); + } + onChanged(); + } + } else { + if (!other.regionLoads_.isEmpty()) { + if (regionLoadsBuilder_.isEmpty()) { + regionLoadsBuilder_.dispose(); + regionLoadsBuilder_ = null; + regionLoads_ = other.regionLoads_; + bitField0_ = (bitField0_ & ~0x00000010); + regionLoadsBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getRegionLoadsFieldBuilder() : null; + } else { + regionLoadsBuilder_.addAllMessages(other.regionLoads_); + } + } + } + if (coprocessorsBuilder_ == null) { + if (!other.coprocessors_.isEmpty()) { + if (coprocessors_.isEmpty()) { + coprocessors_ = other.coprocessors_; + bitField0_ = (bitField0_ & ~0x00000020); + } else { + ensureCoprocessorsIsMutable(); + coprocessors_.addAll(other.coprocessors_); + } + onChanged(); + } + } else { + if (!other.coprocessors_.isEmpty()) { + if (coprocessorsBuilder_.isEmpty()) { + coprocessorsBuilder_.dispose(); + coprocessorsBuilder_ = null; + coprocessors_ = other.coprocessors_; + bitField0_ = (bitField0_ & ~0x00000020); + coprocessorsBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getCoprocessorsFieldBuilder() : null; + } else { + coprocessorsBuilder_.addAllMessages(other.coprocessors_); + } + } + } + if (other.hasReportStartTime()) { + setReportStartTime(other.getReportStartTime()); + } + if (other.hasReportEndTime()) { + setReportEndTime(other.getReportEndTime()); + } + if (other.hasInfoServerPort()) { + setInfoServerPort(other.getInfoServerPort()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + for (int i = 0; i < getRegionLoadsCount(); i++) { + if (!getRegionLoads(i).isInitialized()) { + + return false; + } + } + for (int i = 0; i < getCoprocessorsCount(); i++) { + if (!getCoprocessors(i).isInitialized()) { + + return false; + } + } + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder( + this.getUnknownFields()); + while (true) { + int tag = input.readTag(); + switch (tag) { + case 0: + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + } + break; + } + case 8: { + bitField0_ |= 0x00000001; + numberOfRequests_ = input.readUInt32(); + break; + } + case 16: { + bitField0_ |= 0x00000002; + totalNumberOfRequests_ = input.readUInt32(); + break; + } + case 24: { + bitField0_ |= 0x00000004; + usedHeapMB_ = input.readUInt32(); + break; + } + case 32: { + bitField0_ |= 0x00000008; + maxHeapMB_ = input.readUInt32(); + break; + } + case 42: { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addRegionLoads(subBuilder.buildPartial()); + break; + } + case 50: { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addCoprocessors(subBuilder.buildPartial()); + break; + } + case 56: { + bitField0_ |= 0x00000040; + reportStartTime_ = input.readUInt64(); + break; + } + case 64: { + bitField0_ |= 0x00000080; + reportEndTime_ = input.readUInt64(); + break; + } + case 72: { + bitField0_ |= 0x00000100; + infoServerPort_ = input.readUInt32(); + break; + } + } + } + } + + private int bitField0_; + + // optional uint32 numberOfRequests = 1; + private int numberOfRequests_ ; + public boolean hasNumberOfRequests() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public int getNumberOfRequests() { + return numberOfRequests_; + } + public Builder setNumberOfRequests(int value) { + bitField0_ |= 0x00000001; + numberOfRequests_ = value; + onChanged(); + return this; + } + public Builder clearNumberOfRequests() { + bitField0_ = (bitField0_ & ~0x00000001); + numberOfRequests_ = 0; + onChanged(); + return this; + } + + // optional uint32 totalNumberOfRequests = 2; + private int totalNumberOfRequests_ ; + public boolean hasTotalNumberOfRequests() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public int getTotalNumberOfRequests() { + return totalNumberOfRequests_; + } + public Builder setTotalNumberOfRequests(int value) { + bitField0_ |= 0x00000002; + totalNumberOfRequests_ = value; + onChanged(); + return this; + } + public Builder clearTotalNumberOfRequests() { + bitField0_ = (bitField0_ & ~0x00000002); + totalNumberOfRequests_ = 0; + onChanged(); + return this; + } + + // optional uint32 usedHeapMB = 3; + private int usedHeapMB_ ; + public boolean hasUsedHeapMB() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + public int getUsedHeapMB() { + return usedHeapMB_; + } + public Builder setUsedHeapMB(int value) { + bitField0_ |= 0x00000004; + usedHeapMB_ = value; + onChanged(); + return this; + } + public Builder clearUsedHeapMB() { + bitField0_ = (bitField0_ & ~0x00000004); + usedHeapMB_ = 0; + onChanged(); + return this; + } + + // optional uint32 maxHeapMB = 4; + private int maxHeapMB_ ; + public boolean hasMaxHeapMB() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + public int getMaxHeapMB() { + return maxHeapMB_; + } + public Builder setMaxHeapMB(int value) { + bitField0_ |= 0x00000008; + maxHeapMB_ = value; + onChanged(); + return this; + } + public Builder clearMaxHeapMB() { + bitField0_ = (bitField0_ & ~0x00000008); + maxHeapMB_ = 0; + onChanged(); + return this; + } + + // repeated .RegionLoad regionLoads = 5; + private java.util.List regionLoads_ = + java.util.Collections.emptyList(); + private void ensureRegionLoadsIsMutable() { + if (!((bitField0_ & 0x00000010) == 0x00000010)) { + regionLoads_ = new java.util.ArrayList(regionLoads_); + bitField0_ |= 0x00000010; + } + } + + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder> regionLoadsBuilder_; + + public java.util.List getRegionLoadsList() { + if (regionLoadsBuilder_ == null) { + return java.util.Collections.unmodifiableList(regionLoads_); + } else { + return regionLoadsBuilder_.getMessageList(); + } + } + public int getRegionLoadsCount() { + if (regionLoadsBuilder_ == null) { + return regionLoads_.size(); + } else { + return regionLoadsBuilder_.getCount(); + } + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad getRegionLoads(int index) { + if (regionLoadsBuilder_ == null) { + return regionLoads_.get(index); + } else { + return regionLoadsBuilder_.getMessage(index); + } + } + public Builder setRegionLoads( + int index, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad value) { + if (regionLoadsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureRegionLoadsIsMutable(); + regionLoads_.set(index, value); + onChanged(); + } else { + regionLoadsBuilder_.setMessage(index, value); + } + return this; + } + public Builder setRegionLoads( + int index, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder builderForValue) { + if (regionLoadsBuilder_ == null) { + ensureRegionLoadsIsMutable(); + regionLoads_.set(index, builderForValue.build()); + onChanged(); + } else { + regionLoadsBuilder_.setMessage(index, builderForValue.build()); + } + return this; + } + public Builder addRegionLoads(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad value) { + if (regionLoadsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureRegionLoadsIsMutable(); + regionLoads_.add(value); + onChanged(); + } else { + regionLoadsBuilder_.addMessage(value); + } + return this; + } + public Builder addRegionLoads( + int index, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad value) { + if (regionLoadsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureRegionLoadsIsMutable(); + regionLoads_.add(index, value); + onChanged(); + } else { + regionLoadsBuilder_.addMessage(index, value); + } + return this; + } + public Builder addRegionLoads( + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder builderForValue) { + if (regionLoadsBuilder_ == null) { + ensureRegionLoadsIsMutable(); + regionLoads_.add(builderForValue.build()); + onChanged(); + } else { + regionLoadsBuilder_.addMessage(builderForValue.build()); + } + return this; + } + public Builder addRegionLoads( + int index, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder builderForValue) { + if (regionLoadsBuilder_ == null) { + ensureRegionLoadsIsMutable(); + regionLoads_.add(index, builderForValue.build()); + onChanged(); + } else { + regionLoadsBuilder_.addMessage(index, builderForValue.build()); + } + return this; + } + public Builder addAllRegionLoads( + java.lang.Iterable values) { + if (regionLoadsBuilder_ == null) { + ensureRegionLoadsIsMutable(); + super.addAll(values, regionLoads_); + onChanged(); + } else { + regionLoadsBuilder_.addAllMessages(values); + } + return this; + } + public Builder clearRegionLoads() { + if (regionLoadsBuilder_ == null) { + regionLoads_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000010); + onChanged(); + } else { + regionLoadsBuilder_.clear(); + } + return this; + } + public Builder removeRegionLoads(int index) { + if (regionLoadsBuilder_ == null) { + ensureRegionLoadsIsMutable(); + regionLoads_.remove(index); + onChanged(); + } else { + regionLoadsBuilder_.remove(index); + } + return this; + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder getRegionLoadsBuilder( + int index) { + return getRegionLoadsFieldBuilder().getBuilder(index); + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder getRegionLoadsOrBuilder( + int index) { + if (regionLoadsBuilder_ == null) { + return regionLoads_.get(index); } else { + return regionLoadsBuilder_.getMessageOrBuilder(index); + } + } + public java.util.List + getRegionLoadsOrBuilderList() { + if (regionLoadsBuilder_ != null) { + return regionLoadsBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(regionLoads_); + } + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder addRegionLoadsBuilder() { + return getRegionLoadsFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.getDefaultInstance()); + } + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder addRegionLoadsBuilder( + int index) { + return getRegionLoadsFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.getDefaultInstance()); + } + public java.util.List + getRegionLoadsBuilderList() { + return getRegionLoadsFieldBuilder().getBuilderList(); + } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder> + getRegionLoadsFieldBuilder() { + if (regionLoadsBuilder_ == null) { + regionLoadsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoadOrBuilder>( + regionLoads_, + ((bitField0_ & 0x00000010) == 0x00000010), + getParentForChildren(), + isClean()); + regionLoads_ = null; + } + return regionLoadsBuilder_; + } + + // repeated .Coprocessor coprocessors = 6; + private java.util.List coprocessors_ = + java.util.Collections.emptyList(); + private void ensureCoprocessorsIsMutable() { + if (!((bitField0_ & 0x00000020) == 0x00000020)) { + coprocessors_ = new java.util.ArrayList(coprocessors_); + bitField0_ |= 0x00000020; + } + } + + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder> coprocessorsBuilder_; + + public java.util.List getCoprocessorsList() { + if (coprocessorsBuilder_ == null) { + return java.util.Collections.unmodifiableList(coprocessors_); + } else { + return coprocessorsBuilder_.getMessageList(); + } + } + public int getCoprocessorsCount() { + if (coprocessorsBuilder_ == null) { + return coprocessors_.size(); + } else { + return coprocessorsBuilder_.getCount(); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getCoprocessors(int index) { + if (coprocessorsBuilder_ == null) { + return coprocessors_.get(index); + } else { + return coprocessorsBuilder_.getMessage(index); + } + } + public Builder setCoprocessors( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor value) { + if (coprocessorsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureCoprocessorsIsMutable(); + coprocessors_.set(index, value); + onChanged(); + } else { + coprocessorsBuilder_.setMessage(index, value); + } + return this; + } + public Builder setCoprocessors( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder builderForValue) { + if (coprocessorsBuilder_ == null) { + ensureCoprocessorsIsMutable(); + coprocessors_.set(index, builderForValue.build()); + onChanged(); + } else { + coprocessorsBuilder_.setMessage(index, builderForValue.build()); + } + return this; + } + public Builder addCoprocessors(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor value) { + if (coprocessorsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureCoprocessorsIsMutable(); + coprocessors_.add(value); + onChanged(); + } else { + coprocessorsBuilder_.addMessage(value); + } + return this; + } + public Builder addCoprocessors( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor value) { + if (coprocessorsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureCoprocessorsIsMutable(); + coprocessors_.add(index, value); + onChanged(); + } else { + coprocessorsBuilder_.addMessage(index, value); + } + return this; + } + public Builder addCoprocessors( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder builderForValue) { + if (coprocessorsBuilder_ == null) { + ensureCoprocessorsIsMutable(); + coprocessors_.add(builderForValue.build()); + onChanged(); + } else { + coprocessorsBuilder_.addMessage(builderForValue.build()); + } + return this; + } + public Builder addCoprocessors( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder builderForValue) { + if (coprocessorsBuilder_ == null) { + ensureCoprocessorsIsMutable(); + coprocessors_.add(index, builderForValue.build()); + onChanged(); + } else { + coprocessorsBuilder_.addMessage(index, builderForValue.build()); + } + return this; + } + public Builder addAllCoprocessors( + java.lang.Iterable values) { + if (coprocessorsBuilder_ == null) { + ensureCoprocessorsIsMutable(); + super.addAll(values, coprocessors_); + onChanged(); + } else { + coprocessorsBuilder_.addAllMessages(values); + } + return this; + } + public Builder clearCoprocessors() { + if (coprocessorsBuilder_ == null) { + coprocessors_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000020); + onChanged(); + } else { + coprocessorsBuilder_.clear(); + } + return this; + } + public Builder removeCoprocessors(int index) { + if (coprocessorsBuilder_ == null) { + ensureCoprocessorsIsMutable(); + coprocessors_.remove(index); + onChanged(); + } else { + coprocessorsBuilder_.remove(index); + } + return this; + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder getCoprocessorsBuilder( + int index) { + return getCoprocessorsFieldBuilder().getBuilder(index); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( + int index) { + if (coprocessorsBuilder_ == null) { + return coprocessors_.get(index); } else { + return coprocessorsBuilder_.getMessageOrBuilder(index); + } + } + public java.util.List + getCoprocessorsOrBuilderList() { + if (coprocessorsBuilder_ != null) { + return coprocessorsBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(coprocessors_); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder addCoprocessorsBuilder() { + return getCoprocessorsFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.getDefaultInstance()); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder addCoprocessorsBuilder( + int index) { + return getCoprocessorsFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.getDefaultInstance()); + } + public java.util.List + getCoprocessorsBuilderList() { + return getCoprocessorsFieldBuilder().getBuilderList(); + } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder> + getCoprocessorsFieldBuilder() { + if (coprocessorsBuilder_ == null) { + coprocessorsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder>( + coprocessors_, + ((bitField0_ & 0x00000020) == 0x00000020), + getParentForChildren(), + isClean()); + coprocessors_ = null; + } + return coprocessorsBuilder_; + } + + // optional uint64 reportStartTime = 7; + private long reportStartTime_ ; + public boolean hasReportStartTime() { + return ((bitField0_ & 0x00000040) == 0x00000040); + } + public long getReportStartTime() { + return reportStartTime_; + } + public Builder setReportStartTime(long value) { + bitField0_ |= 0x00000040; + reportStartTime_ = value; + onChanged(); + return this; + } + public Builder clearReportStartTime() { + bitField0_ = (bitField0_ & ~0x00000040); + reportStartTime_ = 0L; + onChanged(); + return this; + } + + // optional uint64 reportEndTime = 8; + private long reportEndTime_ ; + public boolean hasReportEndTime() { + return ((bitField0_ & 0x00000080) == 0x00000080); + } + public long getReportEndTime() { + return reportEndTime_; + } + public Builder setReportEndTime(long value) { + bitField0_ |= 0x00000080; + reportEndTime_ = value; + onChanged(); + return this; + } + public Builder clearReportEndTime() { + bitField0_ = (bitField0_ & ~0x00000080); + reportEndTime_ = 0L; + onChanged(); + return this; + } + + // optional uint32 infoServerPort = 9; + private int infoServerPort_ ; + public boolean hasInfoServerPort() { + return ((bitField0_ & 0x00000100) == 0x00000100); + } + public int getInfoServerPort() { + return infoServerPort_; + } + public Builder setInfoServerPort(int value) { + bitField0_ |= 0x00000100; + infoServerPort_ = value; + onChanged(); + return this; + } + public Builder clearInfoServerPort() { + bitField0_ = (bitField0_ & ~0x00000100); + infoServerPort_ = 0; + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:ServerLoad) + } + + static { + defaultInstance = new ServerLoad(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:ServerLoad) + } + public interface LiveServerInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { @@ -1377,8 +4123,8 @@ public final class ClusterStatusProtos { // required .ServerLoad serverLoad = 2; boolean hasServerLoad(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getServerLoad(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder getServerLoadOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getServerLoad(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getServerLoadOrBuilder(); } public static final class LiveServerInfo extends com.google.protobuf.GeneratedMessage @@ -1424,20 +4170,20 @@ public final class ClusterStatusProtos { // required .ServerLoad serverLoad = 2; public static final int SERVERLOAD_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad serverLoad_; + private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad serverLoad_; public boolean hasServerLoad() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getServerLoad() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getServerLoad() { return serverLoad_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder getServerLoadOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getServerLoadOrBuilder() { return serverLoad_; } private void initFields() { server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); - serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -1665,7 +4411,7 @@ public final class ClusterStatusProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (serverLoadBuilder_ == null) { - serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); } else { serverLoadBuilder_.clear(); } @@ -1803,7 +4549,7 @@ public final class ClusterStatusProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.newBuilder(); if (hasServerLoad()) { subBuilder.mergeFrom(getServerLoad()); } @@ -1908,20 +4654,20 @@ public final class ClusterStatusProtos { } // required .ServerLoad serverLoad = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder> serverLoadBuilder_; + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder> serverLoadBuilder_; public boolean hasServerLoad() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getServerLoad() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getServerLoad() { if (serverLoadBuilder_ == null) { return serverLoad_; } else { return serverLoadBuilder_.getMessage(); } } - public Builder setServerLoad(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad value) { + public Builder setServerLoad(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad value) { if (serverLoadBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -1935,7 +4681,7 @@ public final class ClusterStatusProtos { return this; } public Builder setServerLoad( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder builderForValue) { if (serverLoadBuilder_ == null) { serverLoad_ = builderForValue.build(); onChanged(); @@ -1945,12 +4691,12 @@ public final class ClusterStatusProtos { bitField0_ |= 0x00000002; return this; } - public Builder mergeServerLoad(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad value) { + public Builder mergeServerLoad(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad value) { if (serverLoadBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - serverLoad_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance()) { + serverLoad_ != org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance()) { serverLoad_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.newBuilder(serverLoad_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.newBuilder(serverLoad_).mergeFrom(value).buildPartial(); } else { serverLoad_ = value; } @@ -1963,7 +4709,7 @@ public final class ClusterStatusProtos { } public Builder clearServerLoad() { if (serverLoadBuilder_ == null) { - serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); onChanged(); } else { serverLoadBuilder_.clear(); @@ -1971,12 +4717,12 @@ public final class ClusterStatusProtos { bitField0_ = (bitField0_ & ~0x00000002); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder getServerLoadBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder getServerLoadBuilder() { bitField0_ |= 0x00000002; onChanged(); return getServerLoadFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder getServerLoadOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getServerLoadOrBuilder() { if (serverLoadBuilder_ != null) { return serverLoadBuilder_.getMessageOrBuilder(); } else { @@ -1984,11 +4730,11 @@ public final class ClusterStatusProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder> getServerLoadFieldBuilder() { if (serverLoadBuilder_ == null) { serverLoadBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder>( serverLoad_, getParentForChildren(), isClean()); @@ -4337,6 +7083,16 @@ public final class ClusterStatusProtos { com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_RegionInTransition_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor + internal_static_RegionLoad_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_RegionLoad_fieldAccessorTable; + private static com.google.protobuf.Descriptors.Descriptor + internal_static_ServerLoad_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_ServerLoad_fieldAccessorTable; + private static com.google.protobuf.Descriptors.Descriptor internal_static_LiveServerInfo_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable @@ -4365,19 +7121,37 @@ public final class ClusterStatusProtos { "G\020\007\022\t\n\005SPLIT\020\010\022\017\n\013FAILED_OPEN\020\t\022\020\n\014FAILE" + "D_CLOSE\020\n\"W\n\022RegionInTransition\022\036\n\004spec\030" + "\001 \002(\0132\020.RegionSpecifier\022!\n\013regionState\030\002", - " \002(\0132\014.RegionState\"N\n\016LiveServerInfo\022\033\n\006" + - "server\030\001 \002(\0132\013.ServerName\022\037\n\nserverLoad\030" + - "\002 \002(\0132\013.ServerLoad\"\327\002\n\rClusterStatus\022.\n\014" + - "hbaseVersion\030\001 \001(\0132\030.HBaseVersionFileCon" + - "tent\022$\n\013liveServers\030\002 \003(\0132\017.LiveServerIn" + - "fo\022 \n\013deadServers\030\003 \003(\0132\013.ServerName\0220\n\023" + - "regionsInTransition\030\004 \003(\0132\023.RegionInTran" + - "sition\022\035\n\tclusterId\030\005 \001(\0132\n.ClusterId\022(\n" + - "\022masterCoprocessors\030\006 \003(\0132\014.Coprocessor\022" + - "\033\n\006master\030\007 \001(\0132\013.ServerName\022\"\n\rbackupMa", - "sters\030\010 \003(\0132\013.ServerName\022\022\n\nbalancerOn\030\t" + - " \001(\010BF\n*org.apache.hadoop.hbase.protobuf" + - ".generatedB\023ClusterStatusProtosH\001\240\001\001" + " \002(\0132\014.RegionState\"\260\003\n\nRegionLoad\022)\n\017reg" + + "ionSpecifier\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006" + + "stores\030\002 \001(\r\022\022\n\nstorefiles\030\003 \001(\r\022\037\n\027stor" + + "eUncompressedSizeMB\030\004 \001(\r\022\027\n\017storefileSi" + + "zeMB\030\005 \001(\r\022\026\n\016memstoreSizeMB\030\006 \001(\r\022\034\n\024st" + + "orefileIndexSizeMB\030\007 \001(\r\022\031\n\021readRequests" + + "Count\030\010 \001(\004\022\032\n\022writeRequestsCount\030\t \001(\004\022" + + "\032\n\022totalCompactingKVs\030\n \001(\004\022\033\n\023currentCo" + + "mpactedKVs\030\013 \001(\004\022\027\n\017rootIndexSizeKB\030\014 \001(" + + "\r\022\036\n\026totalStaticIndexSizeKB\030\r \001(\r\022\036\n\026tot", + "alStaticBloomSizeKB\030\016 \001(\r\022\032\n\022completeSeq" + + "uenceId\030\017 \001(\004\"\372\001\n\nServerLoad\022\030\n\020numberOf" + + "Requests\030\001 \001(\r\022\035\n\025totalNumberOfRequests\030" + + "\002 \001(\r\022\022\n\nusedHeapMB\030\003 \001(\r\022\021\n\tmaxHeapMB\030\004" + + " \001(\r\022 \n\013regionLoads\030\005 \003(\0132\013.RegionLoad\022\"" + + "\n\014coprocessors\030\006 \003(\0132\014.Coprocessor\022\027\n\017re" + + "portStartTime\030\007 \001(\004\022\025\n\rreportEndTime\030\010 \001" + + "(\004\022\026\n\016infoServerPort\030\t \001(\r\"N\n\016LiveServer" + + "Info\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\037\n\nser" + + "verLoad\030\002 \002(\0132\013.ServerLoad\"\327\002\n\rClusterSt", + "atus\022.\n\014hbaseVersion\030\001 \001(\0132\030.HBaseVersio" + + "nFileContent\022$\n\013liveServers\030\002 \003(\0132\017.Live" + + "ServerInfo\022 \n\013deadServers\030\003 \003(\0132\013.Server" + + "Name\0220\n\023regionsInTransition\030\004 \003(\0132\023.Regi" + + "onInTransition\022\035\n\tclusterId\030\005 \001(\0132\n.Clus" + + "terId\022(\n\022masterCoprocessors\030\006 \003(\0132\014.Copr" + + "ocessor\022\033\n\006master\030\007 \001(\0132\013.ServerName\022\"\n\r" + + "backupMasters\030\010 \003(\0132\013.ServerName\022\022\n\nbala" + + "ncerOn\030\t \001(\010BF\n*org.apache.hadoop.hbase." + + "protobuf.generatedB\023ClusterStatusProtosH", + "\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -4400,8 +7174,24 @@ public final class ClusterStatusProtos { new java.lang.String[] { "Spec", "RegionState", }, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionInTransition.class, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionInTransition.Builder.class); - internal_static_LiveServerInfo_descriptor = + internal_static_RegionLoad_descriptor = getDescriptor().getMessageTypes().get(2); + internal_static_RegionLoad_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_RegionLoad_descriptor, + new java.lang.String[] { "RegionSpecifier", "Stores", "Storefiles", "StoreUncompressedSizeMB", "StorefileSizeMB", "MemstoreSizeMB", "StorefileIndexSizeMB", "ReadRequestsCount", "WriteRequestsCount", "TotalCompactingKVs", "CurrentCompactedKVs", "RootIndexSizeKB", "TotalStaticIndexSizeKB", "TotalStaticBloomSizeKB", "CompleteSequenceId", }, + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.class, + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad.Builder.class); + internal_static_ServerLoad_descriptor = + getDescriptor().getMessageTypes().get(3); + internal_static_ServerLoad_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_ServerLoad_descriptor, + new java.lang.String[] { "NumberOfRequests", "TotalNumberOfRequests", "UsedHeapMB", "MaxHeapMB", "RegionLoads", "Coprocessors", "ReportStartTime", "ReportEndTime", "InfoServerPort", }, + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.class, + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder.class); + internal_static_LiveServerInfo_descriptor = + getDescriptor().getMessageTypes().get(4); internal_static_LiveServerInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_LiveServerInfo_descriptor, @@ -4409,7 +7199,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.LiveServerInfo.class, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.LiveServerInfo.Builder.class); internal_static_ClusterStatus_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(5); internal_static_ClusterStatus_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ClusterStatus_descriptor, diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java index a2844f6..fba358b 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java @@ -8,6 +8,491 @@ public final class FilterProtos { public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } + public interface FilterOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // required string name = 1; + boolean hasName(); + String getName(); + + // optional bytes serializedFilter = 2; + boolean hasSerializedFilter(); + com.google.protobuf.ByteString getSerializedFilter(); + } + public static final class Filter extends + com.google.protobuf.GeneratedMessage + implements FilterOrBuilder { + // Use Filter.newBuilder() to construct. + private Filter(Builder builder) { + super(builder); + } + private Filter(boolean noInit) {} + + private static final Filter defaultInstance; + public static Filter getDefaultInstance() { + return defaultInstance; + } + + public Filter getDefaultInstanceForType() { + return defaultInstance; + } + + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_Filter_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_Filter_fieldAccessorTable; + } + + private int bitField0_; + // required string name = 1; + public static final int NAME_FIELD_NUMBER = 1; + private java.lang.Object name_; + public boolean hasName() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public String getName() { + java.lang.Object ref = name_; + if (ref instanceof String) { + return (String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + String s = bs.toStringUtf8(); + if (com.google.protobuf.Internal.isValidUtf8(bs)) { + name_ = s; + } + return s; + } + } + private com.google.protobuf.ByteString getNameBytes() { + java.lang.Object ref = name_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((String) ref); + name_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional bytes serializedFilter = 2; + public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString serializedFilter_; + public boolean hasSerializedFilter() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public com.google.protobuf.ByteString getSerializedFilter() { + return serializedFilter_; + } + + private void initFields() { + name_ = ""; + serializedFilter_ = com.google.protobuf.ByteString.EMPTY; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + if (!hasName()) { + memoizedIsInitialized = 0; + return false; + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, getNameBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, serializedFilter_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, getNameBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, serializedFilter_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter other = (org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter) obj; + + boolean result = true; + result = result && (hasName() == other.hasName()); + if (hasName()) { + result = result && getName() + .equals(other.getName()); + } + result = result && (hasSerializedFilter() == other.hasSerializedFilter()); + if (hasSerializedFilter()) { + result = result && getSerializedFilter() + .equals(other.getSerializedFilter()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + @java.lang.Override + public int hashCode() { + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasName()) { + hash = (37 * hash) + NAME_FIELD_NUMBER; + hash = (53 * hash) + getName().hashCode(); + } + if (hasSerializedFilter()) { + hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER; + hash = (53 * hash) + getSerializedFilter().hashCode(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return newBuilder().mergeFrom(data, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom(java.io.InputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + Builder builder = newBuilder(); + if (builder.mergeDelimitedFrom(input, extensionRegistry)) { + return builder.buildParsed(); + } else { + return null; + } + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return newBuilder().mergeFrom(input).buildParsed(); + } + public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return newBuilder().mergeFrom(input, extensionRegistry) + .buildParsed(); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_Filter_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_Filter_fieldAccessorTable; + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + name_ = ""; + bitField0_ = (bitField0_ & ~0x00000001); + serializedFilter_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000002); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDescriptor(); + } + + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter build() { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter buildParsed() + throws com.google.protobuf.InvalidProtocolBufferException { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException( + result).asInvalidProtocolBufferException(); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter result = new org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.name_ = name_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.serializedFilter_ = serializedFilter_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()) return this; + if (other.hasName()) { + setName(other.getName()); + } + if (other.hasSerializedFilter()) { + setSerializedFilter(other.getSerializedFilter()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + if (!hasName()) { + + return false; + } + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder( + this.getUnknownFields()); + while (true) { + int tag = input.readTag(); + switch (tag) { + case 0: + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + this.setUnknownFields(unknownFields.build()); + onChanged(); + return this; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + name_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + serializedFilter_ = input.readBytes(); + break; + } + } + } + } + + private int bitField0_; + + // required string name = 1; + private java.lang.Object name_ = ""; + public boolean hasName() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public String getName() { + java.lang.Object ref = name_; + if (!(ref instanceof String)) { + String s = ((com.google.protobuf.ByteString) ref).toStringUtf8(); + name_ = s; + return s; + } else { + return (String) ref; + } + } + public Builder setName(String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + name_ = value; + onChanged(); + return this; + } + public Builder clearName() { + bitField0_ = (bitField0_ & ~0x00000001); + name_ = getDefaultInstance().getName(); + onChanged(); + return this; + } + void setName(com.google.protobuf.ByteString value) { + bitField0_ |= 0x00000001; + name_ = value; + onChanged(); + } + + // optional bytes serializedFilter = 2; + private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasSerializedFilter() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + public com.google.protobuf.ByteString getSerializedFilter() { + return serializedFilter_; + } + public Builder setSerializedFilter(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + serializedFilter_ = value; + onChanged(); + return this; + } + public Builder clearSerializedFilter() { + bitField0_ = (bitField0_ & ~0x00000002); + serializedFilter_ = getDefaultInstance().getSerializedFilter(); + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:Filter) + } + + static { + defaultInstance = new Filter(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:Filter) + } + public interface ColumnCountGetFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { @@ -3568,13 +4053,13 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterList.Operator getOperator(); // repeated .Filter filters = 2; - java.util.List + java.util.List getFiltersList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilters(int index); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilters(int index); int getFiltersCount(); - java.util.List + java.util.List getFiltersOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFiltersOrBuilder( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFiltersOrBuilder( int index); } public static final class FilterList extends @@ -3687,21 +4172,21 @@ public final class FilterProtos { // repeated .Filter filters = 2; public static final int FILTERS_FIELD_NUMBER = 2; - private java.util.List filters_; - public java.util.List getFiltersList() { + private java.util.List filters_; + public java.util.List getFiltersList() { return filters_; } - public java.util.List + public java.util.List getFiltersOrBuilderList() { return filters_; } public int getFiltersCount() { return filters_.size(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilters(int index) { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilters(int index) { return filters_.get(index); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFiltersOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFiltersOrBuilder( int index) { return filters_.get(index); } @@ -4076,7 +4561,7 @@ public final class FilterProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(); input.readMessage(subBuilder, extensionRegistry); addFilters(subBuilder.buildPartial()); break; @@ -4112,19 +4597,19 @@ public final class FilterProtos { } // repeated .Filter filters = 2; - private java.util.List filters_ = + private java.util.List filters_ = java.util.Collections.emptyList(); private void ensureFiltersIsMutable() { if (!((bitField0_ & 0x00000002) == 0x00000002)) { - filters_ = new java.util.ArrayList(filters_); + filters_ = new java.util.ArrayList(filters_); bitField0_ |= 0x00000002; } } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> filtersBuilder_; + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> filtersBuilder_; - public java.util.List getFiltersList() { + public java.util.List getFiltersList() { if (filtersBuilder_ == null) { return java.util.Collections.unmodifiableList(filters_); } else { @@ -4138,7 +4623,7 @@ public final class FilterProtos { return filtersBuilder_.getCount(); } } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilters(int index) { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilters(int index) { if (filtersBuilder_ == null) { return filters_.get(index); } else { @@ -4146,7 +4631,7 @@ public final class FilterProtos { } } public Builder setFilters( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + int index, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filtersBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -4160,7 +4645,7 @@ public final class FilterProtos { return this; } public Builder setFilters( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filtersBuilder_ == null) { ensureFiltersIsMutable(); filters_.set(index, builderForValue.build()); @@ -4170,7 +4655,7 @@ public final class FilterProtos { } return this; } - public Builder addFilters(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder addFilters(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filtersBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -4184,7 +4669,7 @@ public final class FilterProtos { return this; } public Builder addFilters( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + int index, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filtersBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -4198,7 +4683,7 @@ public final class FilterProtos { return this; } public Builder addFilters( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filtersBuilder_ == null) { ensureFiltersIsMutable(); filters_.add(builderForValue.build()); @@ -4209,7 +4694,7 @@ public final class FilterProtos { return this; } public Builder addFilters( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filtersBuilder_ == null) { ensureFiltersIsMutable(); filters_.add(index, builderForValue.build()); @@ -4220,7 +4705,7 @@ public final class FilterProtos { return this; } public Builder addAllFilters( - java.lang.Iterable values) { + java.lang.Iterable values) { if (filtersBuilder_ == null) { ensureFiltersIsMutable(); super.addAll(values, filters_); @@ -4250,18 +4735,18 @@ public final class FilterProtos { } return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder getFiltersBuilder( + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder getFiltersBuilder( int index) { return getFiltersFieldBuilder().getBuilder(index); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFiltersOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFiltersOrBuilder( int index) { if (filtersBuilder_ == null) { return filters_.get(index); } else { return filtersBuilder_.getMessageOrBuilder(index); } } - public java.util.List + public java.util.List getFiltersOrBuilderList() { if (filtersBuilder_ != null) { return filtersBuilder_.getMessageOrBuilderList(); @@ -4269,25 +4754,25 @@ public final class FilterProtos { return java.util.Collections.unmodifiableList(filters_); } } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder addFiltersBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder addFiltersBuilder() { return getFiltersFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder addFiltersBuilder( + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder addFiltersBuilder( int index) { return getFiltersFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()); + index, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()); } - public java.util.List + public java.util.List getFiltersBuilderList() { return getFiltersFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> getFiltersFieldBuilder() { if (filtersBuilder_ == null) { filtersBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder>( filters_, ((bitField0_ & 0x00000002) == 0x00000002), getParentForChildren(), @@ -4313,8 +4798,8 @@ public final class FilterProtos { // required .Filter filter = 1; boolean hasFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); } public static final class FilterWrapper extends com.google.protobuf.GeneratedMessage @@ -4347,19 +4832,19 @@ public final class FilterProtos { private int bitField0_; // required .Filter filter = 1; public static final int FILTER_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_; + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_; public boolean hasFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { return filter_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { return filter_; } private void initFields() { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -4556,7 +5041,7 @@ public final class FilterProtos { public Builder clear() { super.clear(); if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } else { filterBuilder_.clear(); } @@ -4666,7 +5151,7 @@ public final class FilterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(); if (hasFilter()) { subBuilder.mergeFrom(getFilter()); } @@ -4681,20 +5166,20 @@ public final class FilterProtos { private int bitField0_; // required .Filter filter = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> filterBuilder_; + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> filterBuilder_; public boolean hasFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { if (filterBuilder_ == null) { return filter_; } else { return filterBuilder_.getMessage(); } } - public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -4708,7 +5193,7 @@ public final class FilterProtos { return this; } public Builder setFilter( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filterBuilder_ == null) { filter_ = builderForValue.build(); onChanged(); @@ -4718,12 +5203,12 @@ public final class FilterProtos { bitField0_ |= 0x00000001; return this; } - public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - filter_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()) { + filter_ != org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()) { filter_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); } else { filter_ = value; } @@ -4736,7 +5221,7 @@ public final class FilterProtos { } public Builder clearFilter() { if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); onChanged(); } else { filterBuilder_.clear(); @@ -4744,12 +5229,12 @@ public final class FilterProtos { bitField0_ = (bitField0_ & ~0x00000001); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder getFilterBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder getFilterBuilder() { bitField0_ |= 0x00000001; onChanged(); return getFilterFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { if (filterBuilder_ != null) { return filterBuilder_.getMessageOrBuilder(); } else { @@ -4757,11 +5242,11 @@ public final class FilterProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> getFilterFieldBuilder() { if (filterBuilder_ == null) { filterBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder>( filter_, getParentForChildren(), isClean()); @@ -10635,8 +11120,8 @@ public final class FilterProtos { // required .Filter filter = 1; boolean hasFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); } public static final class SkipFilter extends com.google.protobuf.GeneratedMessage @@ -10669,19 +11154,19 @@ public final class FilterProtos { private int bitField0_; // required .Filter filter = 1; public static final int FILTER_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_; + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_; public boolean hasFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { return filter_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { return filter_; } private void initFields() { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -10878,7 +11363,7 @@ public final class FilterProtos { public Builder clear() { super.clear(); if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } else { filterBuilder_.clear(); } @@ -10988,7 +11473,7 @@ public final class FilterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(); if (hasFilter()) { subBuilder.mergeFrom(getFilter()); } @@ -11003,20 +11488,20 @@ public final class FilterProtos { private int bitField0_; // required .Filter filter = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> filterBuilder_; + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> filterBuilder_; public boolean hasFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { if (filterBuilder_ == null) { return filter_; } else { return filterBuilder_.getMessage(); } } - public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -11030,7 +11515,7 @@ public final class FilterProtos { return this; } public Builder setFilter( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filterBuilder_ == null) { filter_ = builderForValue.build(); onChanged(); @@ -11040,12 +11525,12 @@ public final class FilterProtos { bitField0_ |= 0x00000001; return this; } - public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - filter_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()) { + filter_ != org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()) { filter_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); } else { filter_ = value; } @@ -11058,7 +11543,7 @@ public final class FilterProtos { } public Builder clearFilter() { if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); onChanged(); } else { filterBuilder_.clear(); @@ -11066,12 +11551,12 @@ public final class FilterProtos { bitField0_ = (bitField0_ & ~0x00000001); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder getFilterBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder getFilterBuilder() { bitField0_ |= 0x00000001; onChanged(); return getFilterFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { if (filterBuilder_ != null) { return filterBuilder_.getMessageOrBuilder(); } else { @@ -11079,11 +11564,11 @@ public final class FilterProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> getFilterFieldBuilder() { if (filterBuilder_ == null) { filterBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder>( filter_, getParentForChildren(), isClean()); @@ -12007,8 +12492,8 @@ public final class FilterProtos { // required .Filter filter = 1; boolean hasFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); } public static final class WhileMatchFilter extends com.google.protobuf.GeneratedMessage @@ -12041,19 +12526,19 @@ public final class FilterProtos { private int bitField0_; // required .Filter filter = 1; public static final int FILTER_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_; + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_; public boolean hasFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { return filter_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { return filter_; } private void initFields() { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -12250,7 +12735,7 @@ public final class FilterProtos { public Builder clear() { super.clear(); if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); } else { filterBuilder_.clear(); } @@ -12360,7 +12845,7 @@ public final class FilterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(); if (hasFilter()) { subBuilder.mergeFrom(getFilter()); } @@ -12375,20 +12860,20 @@ public final class FilterProtos { private int bitField0_; // required .Filter filter = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> filterBuilder_; + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> filterBuilder_; public boolean hasFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getFilter() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter() { if (filterBuilder_ == null) { return filter_; } else { return filterBuilder_.getMessage(); } } - public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder setFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -12402,7 +12887,7 @@ public final class FilterProtos { return this; } public Builder setFilter( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder builderForValue) { if (filterBuilder_ == null) { filter_ = builderForValue.build(); onChanged(); @@ -12412,12 +12897,12 @@ public final class FilterProtos { bitField0_ |= 0x00000001; return this; } - public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter value) { + public Builder mergeFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter value) { if (filterBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - filter_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()) { + filter_ != org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance()) { filter_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.newBuilder(filter_).mergeFrom(value).buildPartial(); } else { filter_ = value; } @@ -12430,7 +12915,7 @@ public final class FilterProtos { } public Builder clearFilter() { if (filterBuilder_ == null) { - filter_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + filter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.getDefaultInstance(); onChanged(); } else { filterBuilder_.clear(); @@ -12438,12 +12923,12 @@ public final class FilterProtos { bitField0_ = (bitField0_ & ~0x00000001); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder getFilterBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder getFilterBuilder() { bitField0_ |= 0x00000001; onChanged(); return getFilterFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder getFilterOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder() { if (filterBuilder_ != null) { return filterBuilder_.getMessageOrBuilder(); } else { @@ -12451,11 +12936,11 @@ public final class FilterProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder> getFilterFieldBuilder() { if (filterBuilder_ == null) { filterBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder>( filter_, getParentForChildren(), isClean()); @@ -12476,6 +12961,11 @@ public final class FilterProtos { } private static com.google.protobuf.Descriptors.Descriptor + internal_static_Filter_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_Filter_fieldAccessorTable; + private static com.google.protobuf.Descriptors.Descriptor internal_static_ColumnCountGetFilter_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable @@ -12615,58 +13105,67 @@ public final class FilterProtos { static { java.lang.String[] descriptorData = { "\n\014Filter.proto\032\013hbase.proto\032\020Comparator." + - "proto\"%\n\024ColumnCountGetFilter\022\r\n\005limit\030\001" + - " \002(\005\"M\n\026ColumnPaginationFilter\022\r\n\005limit\030" + - "\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014columnOffset\030\003 " + - "\001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006prefix\030\001 \002(" + - "\014\"q\n\021ColumnRangeFilter\022\021\n\tminColumn\030\001 \001(" + - "\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022\021\n\tmaxColu" + - "mn\030\003 \001(\014\022\032\n\022maxColumnInclusive\030\004 \001(\010\"Q\n\r" + - "CompareFilter\022\037\n\tcompareOp\030\001 \002(\0162\014.Compa" + - "reType\022\037\n\ncomparator\030\002 \001(\0132\013.Comparator\"", - "\212\001\n\025DependentColumnFilter\022%\n\rcompareFilt" + - "er\030\001 \002(\0132\016.CompareFilter\022\024\n\014columnFamily" + - "\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(\014\022\033\n\023dropD" + - "ependentColumn\030\004 \001(\010\"5\n\014FamilyFilter\022%\n\r" + - "compareFilter\030\001 \002(\0132\016.CompareFilter\"\200\001\n\n" + - "FilterList\022&\n\010operator\030\001 \002(\0162\024.FilterLis" + - "t.Operator\022\030\n\007filters\030\002 \003(\0132\007.Filter\"0\n\010" + - "Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n\rMUST_PASS" + - "_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006filter\030\001 \002(\0132" + - "\007.Filter\"\024\n\022FirstKeyOnlyFilter\";\n%FirstK", - "eyValueMatchingQualifiersFilter\022\022\n\nquali" + - "fiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022&\n\rfuzzyKe" + - "ysData\030\001 \003(\0132\017.BytesBytesPair\")\n\023Inclusi" + - "veStopFilter\022\022\n\nstopRowKey\030\001 \001(\014\"!\n\rKeyO" + - "nlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n\032MultipleC" + - "olumnPrefixFilter\022\026\n\016sortedPrefixes\030\001 \003(" + - "\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 \002(\003\"\036\n\014Pre" + - "fixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017QualifierFi" + + "proto\"0\n\006Filter\022\014\n\004name\030\001 \002(\t\022\030\n\020seriali" + + "zedFilter\030\002 \001(\014\"%\n\024ColumnCountGetFilter\022" + + "\r\n\005limit\030\001 \002(\005\"M\n\026ColumnPaginationFilter" + + "\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014colum" + + "nOffset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006p" + + "refix\030\001 \002(\014\"q\n\021ColumnRangeFilter\022\021\n\tminC" + + "olumn\030\001 \001(\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022" + + "\021\n\tmaxColumn\030\003 \001(\014\022\032\n\022maxColumnInclusive" + + "\030\004 \001(\010\"Q\n\rCompareFilter\022\037\n\tcompareOp\030\001 \002", + "(\0162\014.CompareType\022\037\n\ncomparator\030\002 \001(\0132\013.C" + + "omparator\"\212\001\n\025DependentColumnFilter\022%\n\rc" + + "ompareFilter\030\001 \002(\0132\016.CompareFilter\022\024\n\014co" + + "lumnFamily\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(" + + "\014\022\033\n\023dropDependentColumn\030\004 \001(\010\"5\n\014Family" + + "Filter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareF" + + "ilter\"\200\001\n\nFilterList\022&\n\010operator\030\001 \002(\0162\024" + + ".FilterList.Operator\022\030\n\007filters\030\002 \003(\0132\007." + + "Filter\"0\n\010Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n" + + "\rMUST_PASS_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006fil", + "ter\030\001 \002(\0132\007.Filter\"\024\n\022FirstKeyOnlyFilter" + + "\";\n%FirstKeyValueMatchingQualifiersFilte" + + "r\022\022\n\nqualifiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022" + + "&\n\rfuzzyKeysData\030\001 \003(\0132\017.BytesBytesPair\"" + + ")\n\023InclusiveStopFilter\022\022\n\nstopRowKey\030\001 \001" + + "(\014\"!\n\rKeyOnlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n" + + "\032MultipleColumnPrefixFilter\022\026\n\016sortedPre" + + "fixes\030\001 \003(\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 " + + "\002(\003\"\036\n\014PrefixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017Q" + + "ualifierFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.", + "CompareFilter\"!\n\017RandomRowFilter\022\016\n\006chan" + + "ce\030\001 \002(\002\"2\n\tRowFilter\022%\n\rcompareFilter\030\001" + + " \002(\0132\016.CompareFilter\"[\n\036SingleColumnValu" + + "eExcludeFilter\0229\n\027singleColumnValueFilte" + + "r\030\001 \002(\0132\030.SingleColumnValueFilter\"\276\001\n\027Si" + + "ngleColumnValueFilter\022\024\n\014columnFamily\030\001 " + + "\001(\014\022\027\n\017columnQualifier\030\002 \001(\014\022\037\n\tcompareO" + + "p\030\003 \002(\0162\014.CompareType\022\037\n\ncomparator\030\004 \002(" + + "\0132\013.Comparator\022\027\n\017filterIfMissing\030\005 \001(\010\022" + + "\031\n\021latestVersionOnly\030\006 \001(\010\"%\n\nSkipFilter", + "\022\027\n\006filter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsF" + + "ilter\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFi" + "lter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareFil" + - "ter\"!\n\017RandomRowFilter\022\016\n\006chance\030\001 \002(\002\"2", - "\n\tRowFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.Com" + - "pareFilter\"[\n\036SingleColumnValueExcludeFi" + - "lter\0229\n\027singleColumnValueFilter\030\001 \002(\0132\030." + - "SingleColumnValueFilter\"\276\001\n\027SingleColumn" + - "ValueFilter\022\024\n\014columnFamily\030\001 \001(\014\022\027\n\017col" + - "umnQualifier\030\002 \001(\014\022\037\n\tcompareOp\030\003 \002(\0162\014." + - "CompareType\022\037\n\ncomparator\030\004 \002(\0132\013.Compar" + - "ator\022\027\n\017filterIfMissing\030\005 \001(\010\022\031\n\021latestV" + - "ersionOnly\030\006 \001(\010\"%\n\nSkipFilter\022\027\n\006filter" + - "\030\001 \002(\0132\007.Filter\"*\n\020TimestampsFilter\022\026\n\nt", - "imestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFilter\022%\n\rco" + - "mpareFilter\030\001 \002(\0132\016.CompareFilter\"+\n\020Whi" + - "leMatchFilter\022\027\n\006filter\030\001 \002(\0132\007.FilterBB" + - "\n*org.apache.hadoop.hbase.protobuf.gener" + - "atedB\014FilterProtosH\001\210\001\001\240\001\001" + "ter\"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132" + + "\007.FilterBB\n*org.apache.hadoop.hbase.prot" + + "obuf.generatedB\014FilterProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; - internal_static_ColumnCountGetFilter_descriptor = + internal_static_Filter_descriptor = getDescriptor().getMessageTypes().get(0); + internal_static_Filter_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_Filter_descriptor, + new java.lang.String[] { "Name", "SerializedFilter", }, + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.class, + org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter.Builder.class); + internal_static_ColumnCountGetFilter_descriptor = + getDescriptor().getMessageTypes().get(1); internal_static_ColumnCountGetFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ColumnCountGetFilter_descriptor, @@ -12674,7 +13173,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnCountGetFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnCountGetFilter.Builder.class); internal_static_ColumnPaginationFilter_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(2); internal_static_ColumnPaginationFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ColumnPaginationFilter_descriptor, @@ -12682,7 +13181,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnPaginationFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnPaginationFilter.Builder.class); internal_static_ColumnPrefixFilter_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(3); internal_static_ColumnPrefixFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ColumnPrefixFilter_descriptor, @@ -12690,7 +13189,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnPrefixFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnPrefixFilter.Builder.class); internal_static_ColumnRangeFilter_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(4); internal_static_ColumnRangeFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ColumnRangeFilter_descriptor, @@ -12698,7 +13197,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnRangeFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ColumnRangeFilter.Builder.class); internal_static_CompareFilter_descriptor = - getDescriptor().getMessageTypes().get(4); + getDescriptor().getMessageTypes().get(5); internal_static_CompareFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_CompareFilter_descriptor, @@ -12706,7 +13205,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder.class); internal_static_DependentColumnFilter_descriptor = - getDescriptor().getMessageTypes().get(5); + getDescriptor().getMessageTypes().get(6); internal_static_DependentColumnFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_DependentColumnFilter_descriptor, @@ -12714,7 +13213,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.DependentColumnFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.DependentColumnFilter.Builder.class); internal_static_FamilyFilter_descriptor = - getDescriptor().getMessageTypes().get(6); + getDescriptor().getMessageTypes().get(7); internal_static_FamilyFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FamilyFilter_descriptor, @@ -12722,7 +13221,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FamilyFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FamilyFilter.Builder.class); internal_static_FilterList_descriptor = - getDescriptor().getMessageTypes().get(7); + getDescriptor().getMessageTypes().get(8); internal_static_FilterList_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FilterList_descriptor, @@ -12730,7 +13229,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterList.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterList.Builder.class); internal_static_FilterWrapper_descriptor = - getDescriptor().getMessageTypes().get(8); + getDescriptor().getMessageTypes().get(9); internal_static_FilterWrapper_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FilterWrapper_descriptor, @@ -12738,7 +13237,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterWrapper.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterWrapper.Builder.class); internal_static_FirstKeyOnlyFilter_descriptor = - getDescriptor().getMessageTypes().get(9); + getDescriptor().getMessageTypes().get(10); internal_static_FirstKeyOnlyFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FirstKeyOnlyFilter_descriptor, @@ -12746,7 +13245,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FirstKeyOnlyFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FirstKeyOnlyFilter.Builder.class); internal_static_FirstKeyValueMatchingQualifiersFilter_descriptor = - getDescriptor().getMessageTypes().get(10); + getDescriptor().getMessageTypes().get(11); internal_static_FirstKeyValueMatchingQualifiersFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FirstKeyValueMatchingQualifiersFilter_descriptor, @@ -12754,7 +13253,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FirstKeyValueMatchingQualifiersFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FirstKeyValueMatchingQualifiersFilter.Builder.class); internal_static_FuzzyRowFilter_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(12); internal_static_FuzzyRowFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FuzzyRowFilter_descriptor, @@ -12762,7 +13261,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FuzzyRowFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FuzzyRowFilter.Builder.class); internal_static_InclusiveStopFilter_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(13); internal_static_InclusiveStopFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_InclusiveStopFilter_descriptor, @@ -12770,7 +13269,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.InclusiveStopFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.InclusiveStopFilter.Builder.class); internal_static_KeyOnlyFilter_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(14); internal_static_KeyOnlyFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_KeyOnlyFilter_descriptor, @@ -12778,7 +13277,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.KeyOnlyFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.KeyOnlyFilter.Builder.class); internal_static_MultipleColumnPrefixFilter_descriptor = - getDescriptor().getMessageTypes().get(14); + getDescriptor().getMessageTypes().get(15); internal_static_MultipleColumnPrefixFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_MultipleColumnPrefixFilter_descriptor, @@ -12786,7 +13285,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.MultipleColumnPrefixFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.MultipleColumnPrefixFilter.Builder.class); internal_static_PageFilter_descriptor = - getDescriptor().getMessageTypes().get(15); + getDescriptor().getMessageTypes().get(16); internal_static_PageFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_PageFilter_descriptor, @@ -12794,7 +13293,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.PageFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.PageFilter.Builder.class); internal_static_PrefixFilter_descriptor = - getDescriptor().getMessageTypes().get(16); + getDescriptor().getMessageTypes().get(17); internal_static_PrefixFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_PrefixFilter_descriptor, @@ -12802,7 +13301,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.PrefixFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.PrefixFilter.Builder.class); internal_static_QualifierFilter_descriptor = - getDescriptor().getMessageTypes().get(17); + getDescriptor().getMessageTypes().get(18); internal_static_QualifierFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_QualifierFilter_descriptor, @@ -12810,7 +13309,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.QualifierFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.QualifierFilter.Builder.class); internal_static_RandomRowFilter_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(19); internal_static_RandomRowFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RandomRowFilter_descriptor, @@ -12818,7 +13317,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.RandomRowFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.RandomRowFilter.Builder.class); internal_static_RowFilter_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(20); internal_static_RowFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RowFilter_descriptor, @@ -12826,7 +13325,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.RowFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.RowFilter.Builder.class); internal_static_SingleColumnValueExcludeFilter_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(21); internal_static_SingleColumnValueExcludeFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SingleColumnValueExcludeFilter_descriptor, @@ -12834,7 +13333,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueExcludeFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueExcludeFilter.Builder.class); internal_static_SingleColumnValueFilter_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(22); internal_static_SingleColumnValueFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SingleColumnValueFilter_descriptor, @@ -12842,7 +13341,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder.class); internal_static_SkipFilter_descriptor = - getDescriptor().getMessageTypes().get(22); + getDescriptor().getMessageTypes().get(23); internal_static_SkipFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SkipFilter_descriptor, @@ -12850,7 +13349,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SkipFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SkipFilter.Builder.class); internal_static_TimestampsFilter_descriptor = - getDescriptor().getMessageTypes().get(23); + getDescriptor().getMessageTypes().get(24); internal_static_TimestampsFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TimestampsFilter_descriptor, @@ -12858,7 +13357,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.TimestampsFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.TimestampsFilter.Builder.class); internal_static_ValueFilter_descriptor = - getDescriptor().getMessageTypes().get(24); + getDescriptor().getMessageTypes().get(25); internal_static_ValueFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ValueFilter_descriptor, @@ -12866,7 +13365,7 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ValueFilter.class, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.ValueFilter.Builder.class); internal_static_WhileMatchFilter_descriptor = - getDescriptor().getMessageTypes().get(25); + getDescriptor().getMessageTypes().get(26); internal_static_WhileMatchFilter_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_WhileMatchFilter_descriptor, diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java index a566592..1ba0177 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java @@ -8,87 +8,6 @@ public final class HBaseProtos { public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } - public enum CellType - implements com.google.protobuf.ProtocolMessageEnum { - MINIMUM(0, 0), - PUT(1, 4), - DELETE(2, 8), - DELETE_COLUMN(3, 12), - DELETE_FAMILY(4, 14), - MAXIMUM(5, 255), - ; - - public static final int MINIMUM_VALUE = 0; - public static final int PUT_VALUE = 4; - public static final int DELETE_VALUE = 8; - public static final int DELETE_COLUMN_VALUE = 12; - public static final int DELETE_FAMILY_VALUE = 14; - public static final int MAXIMUM_VALUE = 255; - - - public final int getNumber() { return value; } - - public static CellType valueOf(int value) { - switch (value) { - case 0: return MINIMUM; - case 4: return PUT; - case 8: return DELETE; - case 12: return DELETE_COLUMN; - case 14: return DELETE_FAMILY; - case 255: return MAXIMUM; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap() { - public CellType findValueByNumber(int number) { - return CellType.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor().getEnumTypes().get(0); - } - - private static final CellType[] VALUES = { - MINIMUM, PUT, DELETE, DELETE_COLUMN, DELETE_FAMILY, MAXIMUM, - }; - - public static CellType valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private CellType(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:CellType) - } - public enum CompareType implements com.google.protobuf.ProtocolMessageEnum { LESS(0, 0), @@ -146,7 +65,7 @@ public final class HBaseProtos { } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor().getEnumTypes().get(1); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor().getEnumTypes().get(0); } private static final CompareType[] VALUES = { @@ -173,135 +92,174 @@ public final class HBaseProtos { // @@protoc_insertion_point(enum_scope:CompareType) } - public interface CellOrBuilder + public interface TableSchemaOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bytes row = 1; - boolean hasRow(); - com.google.protobuf.ByteString getRow(); - - // optional bytes family = 2; - boolean hasFamily(); - com.google.protobuf.ByteString getFamily(); - - // optional bytes qualifier = 3; - boolean hasQualifier(); - com.google.protobuf.ByteString getQualifier(); + // optional bytes name = 1; + boolean hasName(); + com.google.protobuf.ByteString getName(); - // optional uint64 timestamp = 4; - boolean hasTimestamp(); - long getTimestamp(); + // repeated .BytesBytesPair attributes = 2; + java.util.List + getAttributesList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index); + int getAttributesCount(); + java.util.List + getAttributesOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( + int index); - // optional .CellType cellType = 5; - boolean hasCellType(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType getCellType(); + // repeated .ColumnFamilySchema columnFamilies = 3; + java.util.List + getColumnFamiliesList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index); + int getColumnFamiliesCount(); + java.util.List + getColumnFamiliesOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder( + int index); - // optional bytes value = 6; - boolean hasValue(); - com.google.protobuf.ByteString getValue(); + // repeated .NameStringPair configuration = 4; + java.util.List + getConfigurationList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index); + int getConfigurationCount(); + java.util.List + getConfigurationOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( + int index); } - public static final class Cell extends + public static final class TableSchema extends com.google.protobuf.GeneratedMessage - implements CellOrBuilder { - // Use Cell.newBuilder() to construct. - private Cell(Builder builder) { + implements TableSchemaOrBuilder { + // Use TableSchema.newBuilder() to construct. + private TableSchema(Builder builder) { super(builder); } - private Cell(boolean noInit) {} + private TableSchema(boolean noInit) {} - private static final Cell defaultInstance; - public static Cell getDefaultInstance() { + private static final TableSchema defaultInstance; + public static TableSchema getDefaultInstance() { return defaultInstance; } - public Cell getDefaultInstanceForType() { + public TableSchema getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Cell_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Cell_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_fieldAccessorTable; } private int bitField0_; - // optional bytes row = 1; - public static final int ROW_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString row_; - public boolean hasRow() { + // optional bytes name = 1; + public static final int NAME_FIELD_NUMBER = 1; + private com.google.protobuf.ByteString name_; + public boolean hasName() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public com.google.protobuf.ByteString getRow() { - return row_; + public com.google.protobuf.ByteString getName() { + return name_; } - // optional bytes family = 2; - public static final int FAMILY_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString family_; - public boolean hasFamily() { - return ((bitField0_ & 0x00000002) == 0x00000002); + // repeated .BytesBytesPair attributes = 2; + public static final int ATTRIBUTES_FIELD_NUMBER = 2; + private java.util.List attributes_; + public java.util.List getAttributesList() { + return attributes_; + } + public java.util.List + getAttributesOrBuilderList() { + return attributes_; } - public com.google.protobuf.ByteString getFamily() { - return family_; + public int getAttributesCount() { + return attributes_.size(); } - - // optional bytes qualifier = 3; - public static final int QUALIFIER_FIELD_NUMBER = 3; - private com.google.protobuf.ByteString qualifier_; - public boolean hasQualifier() { - return ((bitField0_ & 0x00000004) == 0x00000004); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { + return attributes_.get(index); } - public com.google.protobuf.ByteString getQualifier() { - return qualifier_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( + int index) { + return attributes_.get(index); } - // optional uint64 timestamp = 4; - public static final int TIMESTAMP_FIELD_NUMBER = 4; - private long timestamp_; - public boolean hasTimestamp() { - return ((bitField0_ & 0x00000008) == 0x00000008); + // repeated .ColumnFamilySchema columnFamilies = 3; + public static final int COLUMNFAMILIES_FIELD_NUMBER = 3; + private java.util.List columnFamilies_; + public java.util.List getColumnFamiliesList() { + return columnFamilies_; } - public long getTimestamp() { - return timestamp_; + public java.util.List + getColumnFamiliesOrBuilderList() { + return columnFamilies_; } - - // optional .CellType cellType = 5; - public static final int CELLTYPE_FIELD_NUMBER = 5; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType cellType_; - public boolean hasCellType() { - return ((bitField0_ & 0x00000010) == 0x00000010); + public int getColumnFamiliesCount() { + return columnFamilies_.size(); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index) { + return columnFamilies_.get(index); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType getCellType() { - return cellType_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder( + int index) { + return columnFamilies_.get(index); } - // optional bytes value = 6; - public static final int VALUE_FIELD_NUMBER = 6; - private com.google.protobuf.ByteString value_; - public boolean hasValue() { - return ((bitField0_ & 0x00000020) == 0x00000020); + // repeated .NameStringPair configuration = 4; + public static final int CONFIGURATION_FIELD_NUMBER = 4; + private java.util.List configuration_; + public java.util.List getConfigurationList() { + return configuration_; } - public com.google.protobuf.ByteString getValue() { - return value_; + public java.util.List + getConfigurationOrBuilderList() { + return configuration_; + } + public int getConfigurationCount() { + return configuration_.size(); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { + return configuration_.get(index); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( + int index) { + return configuration_.get(index); } private void initFields() { - row_ = com.google.protobuf.ByteString.EMPTY; - family_ = com.google.protobuf.ByteString.EMPTY; - qualifier_ = com.google.protobuf.ByteString.EMPTY; - timestamp_ = 0L; - cellType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - value_ = com.google.protobuf.ByteString.EMPTY; + name_ = com.google.protobuf.ByteString.EMPTY; + attributes_ = java.util.Collections.emptyList(); + columnFamilies_ = java.util.Collections.emptyList(); + configuration_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; + for (int i = 0; i < getAttributesCount(); i++) { + if (!getAttributes(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } + for (int i = 0; i < getColumnFamiliesCount(); i++) { + if (!getColumnFamilies(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } + for (int i = 0; i < getConfigurationCount(); i++) { + if (!getConfiguration(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } memoizedIsInitialized = 1; return true; } @@ -310,22 +268,16 @@ public final class HBaseProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, row_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, family_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, qualifier_); + output.writeBytes(1, name_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeUInt64(4, timestamp_); + for (int i = 0; i < attributes_.size(); i++) { + output.writeMessage(2, attributes_.get(i)); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeEnum(5, cellType_.getNumber()); + for (int i = 0; i < columnFamilies_.size(); i++) { + output.writeMessage(3, columnFamilies_.get(i)); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, value_); + for (int i = 0; i < configuration_.size(); i++) { + output.writeMessage(4, configuration_.get(i)); } getUnknownFields().writeTo(output); } @@ -338,27 +290,19 @@ public final class HBaseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, row_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, family_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, qualifier_); + .computeBytesSize(1, name_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + for (int i = 0; i < attributes_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(4, timestamp_); + .computeMessageSize(2, attributes_.get(i)); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + for (int i = 0; i < columnFamilies_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeEnumSize(5, cellType_.getNumber()); + .computeMessageSize(3, columnFamilies_.get(i)); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { + for (int i = 0; i < configuration_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, value_); + .computeMessageSize(4, configuration_.get(i)); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -377,42 +321,23 @@ public final class HBaseProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell) obj; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema) obj; boolean result = true; - result = result && (hasRow() == other.hasRow()); - if (hasRow()) { - result = result && getRow() - .equals(other.getRow()); - } - result = result && (hasFamily() == other.hasFamily()); - if (hasFamily()) { - result = result && getFamily() - .equals(other.getFamily()); - } - result = result && (hasQualifier() == other.hasQualifier()); - if (hasQualifier()) { - result = result && getQualifier() - .equals(other.getQualifier()); - } - result = result && (hasTimestamp() == other.hasTimestamp()); - if (hasTimestamp()) { - result = result && (getTimestamp() - == other.getTimestamp()); - } - result = result && (hasCellType() == other.hasCellType()); - if (hasCellType()) { - result = result && - (getCellType() == other.getCellType()); - } - result = result && (hasValue() == other.hasValue()); - if (hasValue()) { - result = result && getValue() - .equals(other.getValue()); + result = result && (hasName() == other.hasName()); + if (hasName()) { + result = result && getName() + .equals(other.getName()); } + result = result && getAttributesList() + .equals(other.getAttributesList()); + result = result && getColumnFamiliesList() + .equals(other.getColumnFamiliesList()); + result = result && getConfigurationList() + .equals(other.getConfigurationList()); result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -422,69 +347,61 @@ public final class HBaseProtos { public int hashCode() { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasRow()) { - hash = (37 * hash) + ROW_FIELD_NUMBER; - hash = (53 * hash) + getRow().hashCode(); - } - if (hasFamily()) { - hash = (37 * hash) + FAMILY_FIELD_NUMBER; - hash = (53 * hash) + getFamily().hashCode(); - } - if (hasQualifier()) { - hash = (37 * hash) + QUALIFIER_FIELD_NUMBER; - hash = (53 * hash) + getQualifier().hashCode(); + if (hasName()) { + hash = (37 * hash) + NAME_FIELD_NUMBER; + hash = (53 * hash) + getName().hashCode(); } - if (hasTimestamp()) { - hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getTimestamp()); + if (getAttributesCount() > 0) { + hash = (37 * hash) + ATTRIBUTES_FIELD_NUMBER; + hash = (53 * hash) + getAttributesList().hashCode(); } - if (hasCellType()) { - hash = (37 * hash) + CELLTYPE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getCellType()); + if (getColumnFamiliesCount() > 0) { + hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; + hash = (53 * hash) + getColumnFamiliesList().hashCode(); } - if (hasValue()) { - hash = (37 * hash) + VALUE_FIELD_NUMBER; - hash = (53 * hash) + getValue().hashCode(); + if (getConfigurationCount() > 0) { + hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER; + hash = (53 * hash) + getConfigurationList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -493,7 +410,7 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -504,12 +421,12 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -519,7 +436,7 @@ public final class HBaseProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -532,18 +449,18 @@ public final class HBaseProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Cell_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Cell_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -554,6 +471,9 @@ public final class HBaseProtos { } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + getAttributesFieldBuilder(); + getColumnFamiliesFieldBuilder(); + getConfigurationFieldBuilder(); } } private static Builder create() { @@ -562,18 +482,26 @@ public final class HBaseProtos { public Builder clear() { super.clear(); - row_ = com.google.protobuf.ByteString.EMPTY; + name_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); - family_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000002); - qualifier_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000004); - timestamp_ = 0L; - bitField0_ = (bitField0_ & ~0x00000008); - cellType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - bitField0_ = (bitField0_ & ~0x00000010); - value_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000020); + if (attributesBuilder_ == null) { + attributes_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000002); + } else { + attributesBuilder_.clear(); + } + if (columnFamiliesBuilder_ == null) { + columnFamilies_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000004); + } else { + columnFamiliesBuilder_.clear(); + } + if (configurationBuilder_ == null) { + configuration_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000008); + } else { + configurationBuilder_.clear(); + } return this; } @@ -583,24 +511,24 @@ public final class HBaseProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -608,73 +536,161 @@ public final class HBaseProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell(this); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.row_ = row_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.family_ = family_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.qualifier_ = qualifier_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; + result.name_ = name_; + if (attributesBuilder_ == null) { + if (((bitField0_ & 0x00000002) == 0x00000002)) { + attributes_ = java.util.Collections.unmodifiableList(attributes_); + bitField0_ = (bitField0_ & ~0x00000002); + } + result.attributes_ = attributes_; + } else { + result.attributes_ = attributesBuilder_.build(); } - result.timestamp_ = timestamp_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; + if (columnFamiliesBuilder_ == null) { + if (((bitField0_ & 0x00000004) == 0x00000004)) { + columnFamilies_ = java.util.Collections.unmodifiableList(columnFamilies_); + bitField0_ = (bitField0_ & ~0x00000004); + } + result.columnFamilies_ = columnFamilies_; + } else { + result.columnFamilies_ = columnFamiliesBuilder_.build(); } - result.cellType_ = cellType_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { - to_bitField0_ |= 0x00000020; + if (configurationBuilder_ == null) { + if (((bitField0_ & 0x00000008) == 0x00000008)) { + configuration_ = java.util.Collections.unmodifiableList(configuration_); + bitField0_ = (bitField0_ & ~0x00000008); + } + result.configuration_ = configuration_; + } else { + result.configuration_ = configurationBuilder_.build(); } - result.value_ = value_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.getDefaultInstance()) return this; - if (other.hasRow()) { - setRow(other.getRow()); - } - if (other.hasFamily()) { - setFamily(other.getFamily()); - } - if (other.hasQualifier()) { - setQualifier(other.getQualifier()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()) return this; + if (other.hasName()) { + setName(other.getName()); } - if (other.hasTimestamp()) { - setTimestamp(other.getTimestamp()); + if (attributesBuilder_ == null) { + if (!other.attributes_.isEmpty()) { + if (attributes_.isEmpty()) { + attributes_ = other.attributes_; + bitField0_ = (bitField0_ & ~0x00000002); + } else { + ensureAttributesIsMutable(); + attributes_.addAll(other.attributes_); + } + onChanged(); + } + } else { + if (!other.attributes_.isEmpty()) { + if (attributesBuilder_.isEmpty()) { + attributesBuilder_.dispose(); + attributesBuilder_ = null; + attributes_ = other.attributes_; + bitField0_ = (bitField0_ & ~0x00000002); + attributesBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getAttributesFieldBuilder() : null; + } else { + attributesBuilder_.addAllMessages(other.attributes_); + } + } } - if (other.hasCellType()) { - setCellType(other.getCellType()); + if (columnFamiliesBuilder_ == null) { + if (!other.columnFamilies_.isEmpty()) { + if (columnFamilies_.isEmpty()) { + columnFamilies_ = other.columnFamilies_; + bitField0_ = (bitField0_ & ~0x00000004); + } else { + ensureColumnFamiliesIsMutable(); + columnFamilies_.addAll(other.columnFamilies_); + } + onChanged(); + } + } else { + if (!other.columnFamilies_.isEmpty()) { + if (columnFamiliesBuilder_.isEmpty()) { + columnFamiliesBuilder_.dispose(); + columnFamiliesBuilder_ = null; + columnFamilies_ = other.columnFamilies_; + bitField0_ = (bitField0_ & ~0x00000004); + columnFamiliesBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getColumnFamiliesFieldBuilder() : null; + } else { + columnFamiliesBuilder_.addAllMessages(other.columnFamilies_); + } + } } - if (other.hasValue()) { - setValue(other.getValue()); + if (configurationBuilder_ == null) { + if (!other.configuration_.isEmpty()) { + if (configuration_.isEmpty()) { + configuration_ = other.configuration_; + bitField0_ = (bitField0_ & ~0x00000008); + } else { + ensureConfigurationIsMutable(); + configuration_.addAll(other.configuration_); + } + onChanged(); + } + } else { + if (!other.configuration_.isEmpty()) { + if (configurationBuilder_.isEmpty()) { + configurationBuilder_.dispose(); + configurationBuilder_ = null; + configuration_ = other.configuration_; + bitField0_ = (bitField0_ & ~0x00000008); + configurationBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getConfigurationFieldBuilder() : null; + } else { + configurationBuilder_.addAllMessages(other.configuration_); + } + } } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { + for (int i = 0; i < getAttributesCount(); i++) { + if (!getAttributes(i).isInitialized()) { + + return false; + } + } + for (int i = 0; i < getColumnFamiliesCount(); i++) { + if (!getColumnFamilies(i).isInitialized()) { + + return false; + } + } + for (int i = 0; i < getConfigurationCount(); i++) { + if (!getConfiguration(i).isInitialized()) { + + return false; + } + } return true; } @@ -703,38 +719,25 @@ public final class HBaseProtos { } case 10: { bitField0_ |= 0x00000001; - row_ = input.readBytes(); + name_ = input.readBytes(); break; } case 18: { - bitField0_ |= 0x00000002; - family_ = input.readBytes(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addAttributes(subBuilder.buildPartial()); break; } case 26: { - bitField0_ |= 0x00000004; - qualifier_ = input.readBytes(); - break; - } - case 32: { - bitField0_ |= 0x00000008; - timestamp_ = input.readUInt64(); - break; - } - case 40: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType value = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(5, rawValue); - } else { - bitField0_ |= 0x00000010; - cellType_ = value; - } + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addColumnFamilies(subBuilder.buildPartial()); break; } - case 50: { - bitField0_ |= 0x00000020; - value_ = input.readBytes(); + case 34: { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addConfiguration(subBuilder.buildPartial()); break; } } @@ -743,4563 +746,732 @@ public final class HBaseProtos { private int bitField0_; - // optional bytes row = 1; - private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasRow() { + // optional bytes name = 1; + private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasName() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public com.google.protobuf.ByteString getRow() { - return row_; + public com.google.protobuf.ByteString getName() { + return name_; } - public Builder setRow(com.google.protobuf.ByteString value) { + public Builder setName(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; - row_ = value; + name_ = value; onChanged(); return this; } - public Builder clearRow() { + public Builder clearName() { bitField0_ = (bitField0_ & ~0x00000001); - row_ = getDefaultInstance().getRow(); + name_ = getDefaultInstance().getName(); onChanged(); return this; } - // optional bytes family = 2; - private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasFamily() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public com.google.protobuf.ByteString getFamily() { - return family_; - } - public Builder setFamily(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - family_ = value; - onChanged(); - return this; - } - public Builder clearFamily() { - bitField0_ = (bitField0_ & ~0x00000002); - family_ = getDefaultInstance().getFamily(); - onChanged(); - return this; + // repeated .BytesBytesPair attributes = 2; + private java.util.List attributes_ = + java.util.Collections.emptyList(); + private void ensureAttributesIsMutable() { + if (!((bitField0_ & 0x00000002) == 0x00000002)) { + attributes_ = new java.util.ArrayList(attributes_); + bitField0_ |= 0x00000002; + } } - // optional bytes qualifier = 3; - private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasQualifier() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public com.google.protobuf.ByteString getQualifier() { - return qualifier_; + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> attributesBuilder_; + + public java.util.List getAttributesList() { + if (attributesBuilder_ == null) { + return java.util.Collections.unmodifiableList(attributes_); + } else { + return attributesBuilder_.getMessageList(); + } } - public Builder setQualifier(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000004; - qualifier_ = value; - onChanged(); - return this; + public int getAttributesCount() { + if (attributesBuilder_ == null) { + return attributes_.size(); + } else { + return attributesBuilder_.getCount(); + } } - public Builder clearQualifier() { - bitField0_ = (bitField0_ & ~0x00000004); - qualifier_ = getDefaultInstance().getQualifier(); - onChanged(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { + if (attributesBuilder_ == null) { + return attributes_.get(index); + } else { + return attributesBuilder_.getMessage(index); + } + } + public Builder setAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { + if (attributesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureAttributesIsMutable(); + attributes_.set(index, value); + onChanged(); + } else { + attributesBuilder_.setMessage(index, value); + } return this; } - - // optional uint64 timestamp = 4; - private long timestamp_ ; - public boolean hasTimestamp() { - return ((bitField0_ & 0x00000008) == 0x00000008); + public Builder setAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.set(index, builderForValue.build()); + onChanged(); + } else { + attributesBuilder_.setMessage(index, builderForValue.build()); + } + return this; } - public long getTimestamp() { - return timestamp_; + public Builder addAttributes(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { + if (attributesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureAttributesIsMutable(); + attributes_.add(value); + onChanged(); + } else { + attributesBuilder_.addMessage(value); + } + return this; } - public Builder setTimestamp(long value) { - bitField0_ |= 0x00000008; - timestamp_ = value; - onChanged(); + public Builder addAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { + if (attributesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureAttributesIsMutable(); + attributes_.add(index, value); + onChanged(); + } else { + attributesBuilder_.addMessage(index, value); + } return this; } - public Builder clearTimestamp() { - bitField0_ = (bitField0_ & ~0x00000008); - timestamp_ = 0L; - onChanged(); + public Builder addAttributes( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.add(builderForValue.build()); + onChanged(); + } else { + attributesBuilder_.addMessage(builderForValue.build()); + } return this; } - - // optional .CellType cellType = 5; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - public boolean hasCellType() { - return ((bitField0_ & 0x00000010) == 0x00000010); + public Builder addAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.add(index, builderForValue.build()); + onChanged(); + } else { + attributesBuilder_.addMessage(index, builderForValue.build()); + } + return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType getCellType() { - return cellType_; + public Builder addAllAttributes( + java.lang.Iterable values) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + super.addAll(values, attributes_); + onChanged(); + } else { + attributesBuilder_.addAllMessages(values); + } + return this; } - public Builder setCellType(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType value) { - if (value == null) { - throw new NullPointerException(); + public Builder clearAttributes() { + if (attributesBuilder_ == null) { + attributes_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000002); + onChanged(); + } else { + attributesBuilder_.clear(); } - bitField0_ |= 0x00000010; - cellType_ = value; - onChanged(); return this; } - public Builder clearCellType() { - bitField0_ = (bitField0_ & ~0x00000010); - cellType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - onChanged(); + public Builder removeAttributes(int index) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.remove(index); + onChanged(); + } else { + attributesBuilder_.remove(index); + } return this; } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder getAttributesBuilder( + int index) { + return getAttributesFieldBuilder().getBuilder(index); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( + int index) { + if (attributesBuilder_ == null) { + return attributes_.get(index); } else { + return attributesBuilder_.getMessageOrBuilder(index); + } + } + public java.util.List + getAttributesOrBuilderList() { + if (attributesBuilder_ != null) { + return attributesBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(attributes_); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder() { + return getAttributesFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder( + int index) { + return getAttributesFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); + } + public java.util.List + getAttributesBuilderList() { + return getAttributesFieldBuilder().getBuilderList(); + } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> + getAttributesFieldBuilder() { + if (attributesBuilder_ == null) { + attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>( + attributes_, + ((bitField0_ & 0x00000002) == 0x00000002), + getParentForChildren(), + isClean()); + attributes_ = null; + } + return attributesBuilder_; + } - // optional bytes value = 6; - private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasValue() { - return ((bitField0_ & 0x00000020) == 0x00000020); + // repeated .ColumnFamilySchema columnFamilies = 3; + private java.util.List columnFamilies_ = + java.util.Collections.emptyList(); + private void ensureColumnFamiliesIsMutable() { + if (!((bitField0_ & 0x00000004) == 0x00000004)) { + columnFamilies_ = new java.util.ArrayList(columnFamilies_); + bitField0_ |= 0x00000004; + } } - public com.google.protobuf.ByteString getValue() { - return value_; + + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; + + public java.util.List getColumnFamiliesList() { + if (columnFamiliesBuilder_ == null) { + return java.util.Collections.unmodifiableList(columnFamilies_); + } else { + return columnFamiliesBuilder_.getMessageList(); + } } - public Builder setValue(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000020; - value_ = value; - onChanged(); + public int getColumnFamiliesCount() { + if (columnFamiliesBuilder_ == null) { + return columnFamilies_.size(); + } else { + return columnFamiliesBuilder_.getCount(); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index) { + if (columnFamiliesBuilder_ == null) { + return columnFamilies_.get(index); + } else { + return columnFamiliesBuilder_.getMessage(index); + } + } + public Builder setColumnFamilies( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { + if (columnFamiliesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureColumnFamiliesIsMutable(); + columnFamilies_.set(index, value); + onChanged(); + } else { + columnFamiliesBuilder_.setMessage(index, value); + } return this; } - public Builder clearValue() { - bitField0_ = (bitField0_ & ~0x00000020); - value_ = getDefaultInstance().getValue(); - onChanged(); + public Builder setColumnFamilies( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { + if (columnFamiliesBuilder_ == null) { + ensureColumnFamiliesIsMutable(); + columnFamilies_.set(index, builderForValue.build()); + onChanged(); + } else { + columnFamiliesBuilder_.setMessage(index, builderForValue.build()); + } return this; } - - // @@protoc_insertion_point(builder_scope:Cell) - } - - static { - defaultInstance = new Cell(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:Cell) - } - - public interface TableSchemaOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // optional bytes name = 1; - boolean hasName(); - com.google.protobuf.ByteString getName(); - - // repeated .BytesBytesPair attributes = 2; - java.util.List - getAttributesList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index); - int getAttributesCount(); - java.util.List - getAttributesOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( - int index); - - // repeated .ColumnFamilySchema columnFamilies = 3; - java.util.List - getColumnFamiliesList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index); - int getColumnFamiliesCount(); - java.util.List - getColumnFamiliesOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder( - int index); - - // repeated .NameStringPair configuration = 4; - java.util.List - getConfigurationList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index); - int getConfigurationCount(); - java.util.List - getConfigurationOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( - int index); - } - public static final class TableSchema extends - com.google.protobuf.GeneratedMessage - implements TableSchemaOrBuilder { - // Use TableSchema.newBuilder() to construct. - private TableSchema(Builder builder) { - super(builder); - } - private TableSchema(boolean noInit) {} - - private static final TableSchema defaultInstance; - public static TableSchema getDefaultInstance() { - return defaultInstance; - } - - public TableSchema getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_fieldAccessorTable; - } - - private int bitField0_; - // optional bytes name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString name_; - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public com.google.protobuf.ByteString getName() { - return name_; - } - - // repeated .BytesBytesPair attributes = 2; - public static final int ATTRIBUTES_FIELD_NUMBER = 2; - private java.util.List attributes_; - public java.util.List getAttributesList() { - return attributes_; - } - public java.util.List - getAttributesOrBuilderList() { - return attributes_; - } - public int getAttributesCount() { - return attributes_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { - return attributes_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( - int index) { - return attributes_.get(index); - } - - // repeated .ColumnFamilySchema columnFamilies = 3; - public static final int COLUMNFAMILIES_FIELD_NUMBER = 3; - private java.util.List columnFamilies_; - public java.util.List getColumnFamiliesList() { - return columnFamilies_; - } - public java.util.List - getColumnFamiliesOrBuilderList() { - return columnFamilies_; - } - public int getColumnFamiliesCount() { - return columnFamilies_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index) { - return columnFamilies_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder( - int index) { - return columnFamilies_.get(index); - } - - // repeated .NameStringPair configuration = 4; - public static final int CONFIGURATION_FIELD_NUMBER = 4; - private java.util.List configuration_; - public java.util.List getConfigurationList() { - return configuration_; - } - public java.util.List - getConfigurationOrBuilderList() { - return configuration_; - } - public int getConfigurationCount() { - return configuration_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { - return configuration_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( - int index) { - return configuration_.get(index); - } - - private void initFields() { - name_ = com.google.protobuf.ByteString.EMPTY; - attributes_ = java.util.Collections.emptyList(); - columnFamilies_ = java.util.Collections.emptyList(); - configuration_ = java.util.Collections.emptyList(); - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - for (int i = 0; i < getAttributesCount(); i++) { - if (!getAttributes(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; + public Builder addColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { + if (columnFamiliesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureColumnFamiliesIsMutable(); + columnFamilies_.add(value); + onChanged(); + } else { + columnFamiliesBuilder_.addMessage(value); } + return this; } - for (int i = 0; i < getColumnFamiliesCount(); i++) { - if (!getColumnFamilies(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; + public Builder addColumnFamilies( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { + if (columnFamiliesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureColumnFamiliesIsMutable(); + columnFamilies_.add(index, value); + onChanged(); + } else { + columnFamiliesBuilder_.addMessage(index, value); } + return this; } - for (int i = 0; i < getConfigurationCount(); i++) { - if (!getConfiguration(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; + public Builder addColumnFamilies( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { + if (columnFamiliesBuilder_ == null) { + ensureColumnFamiliesIsMutable(); + columnFamilies_.add(builderForValue.build()); + onChanged(); + } else { + columnFamiliesBuilder_.addMessage(builderForValue.build()); } + return this; } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, name_); - } - for (int i = 0; i < attributes_.size(); i++) { - output.writeMessage(2, attributes_.get(i)); - } - for (int i = 0; i < columnFamilies_.size(); i++) { - output.writeMessage(3, columnFamilies_.get(i)); - } - for (int i = 0; i < configuration_.size(); i++) { - output.writeMessage(4, configuration_.get(i)); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, name_); - } - for (int i = 0; i < attributes_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, attributes_.get(i)); - } - for (int i = 0; i < columnFamilies_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(3, columnFamilies_.get(i)); + public Builder addColumnFamilies( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { + if (columnFamiliesBuilder_ == null) { + ensureColumnFamiliesIsMutable(); + columnFamilies_.add(index, builderForValue.build()); + onChanged(); + } else { + columnFamiliesBuilder_.addMessage(index, builderForValue.build()); + } + return this; } - for (int i = 0; i < configuration_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(4, configuration_.get(i)); + public Builder addAllColumnFamilies( + java.lang.Iterable values) { + if (columnFamiliesBuilder_ == null) { + ensureColumnFamiliesIsMutable(); + super.addAll(values, columnFamilies_); + onChanged(); + } else { + columnFamiliesBuilder_.addAllMessages(values); + } + return this; } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema) obj; - - boolean result = true; - result = result && (hasName() == other.hasName()); - if (hasName()) { - result = result && getName() - .equals(other.getName()); - } - result = result && getAttributesList() - .equals(other.getAttributesList()); - result = result && getColumnFamiliesList() - .equals(other.getColumnFamiliesList()); - result = result && getConfigurationList() - .equals(other.getConfigurationList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - @java.lang.Override - public int hashCode() { - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasName()) { - hash = (37 * hash) + NAME_FIELD_NUMBER; - hash = (53 * hash) + getName().hashCode(); - } - if (getAttributesCount() > 0) { - hash = (37 * hash) + ATTRIBUTES_FIELD_NUMBER; - hash = (53 * hash) + getAttributesList().hashCode(); - } - if (getColumnFamiliesCount() > 0) { - hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; - hash = (53 * hash) + getColumnFamiliesList().hashCode(); - } - if (getConfigurationCount() > 0) { - hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER; - hash = (53 * hash) + getConfigurationList().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableSchema_fieldAccessorTable; - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder(BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getAttributesFieldBuilder(); - getColumnFamiliesFieldBuilder(); - getConfigurationFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - name_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000001); - if (attributesBuilder_ == null) { - attributes_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000002); - } else { - attributesBuilder_.clear(); - } - if (columnFamiliesBuilder_ == null) { - columnFamilies_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000004); - } else { - columnFamiliesBuilder_.clear(); - } - if (configurationBuilder_ == null) { - configuration_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000008); - } else { - configurationBuilder_.clear(); - } - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDescriptor(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.name_ = name_; - if (attributesBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002)) { - attributes_ = java.util.Collections.unmodifiableList(attributes_); - bitField0_ = (bitField0_ & ~0x00000002); - } - result.attributes_ = attributes_; - } else { - result.attributes_ = attributesBuilder_.build(); - } - if (columnFamiliesBuilder_ == null) { - if (((bitField0_ & 0x00000004) == 0x00000004)) { - columnFamilies_ = java.util.Collections.unmodifiableList(columnFamilies_); - bitField0_ = (bitField0_ & ~0x00000004); - } - result.columnFamilies_ = columnFamilies_; - } else { - result.columnFamilies_ = columnFamiliesBuilder_.build(); - } - if (configurationBuilder_ == null) { - if (((bitField0_ & 0x00000008) == 0x00000008)) { - configuration_ = java.util.Collections.unmodifiableList(configuration_); - bitField0_ = (bitField0_ & ~0x00000008); - } - result.configuration_ = configuration_; - } else { - result.configuration_ = configurationBuilder_.build(); - } - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()) return this; - if (other.hasName()) { - setName(other.getName()); - } - if (attributesBuilder_ == null) { - if (!other.attributes_.isEmpty()) { - if (attributes_.isEmpty()) { - attributes_ = other.attributes_; - bitField0_ = (bitField0_ & ~0x00000002); - } else { - ensureAttributesIsMutable(); - attributes_.addAll(other.attributes_); - } - onChanged(); - } - } else { - if (!other.attributes_.isEmpty()) { - if (attributesBuilder_.isEmpty()) { - attributesBuilder_.dispose(); - attributesBuilder_ = null; - attributes_ = other.attributes_; - bitField0_ = (bitField0_ & ~0x00000002); - attributesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getAttributesFieldBuilder() : null; - } else { - attributesBuilder_.addAllMessages(other.attributes_); - } - } - } - if (columnFamiliesBuilder_ == null) { - if (!other.columnFamilies_.isEmpty()) { - if (columnFamilies_.isEmpty()) { - columnFamilies_ = other.columnFamilies_; - bitField0_ = (bitField0_ & ~0x00000004); - } else { - ensureColumnFamiliesIsMutable(); - columnFamilies_.addAll(other.columnFamilies_); - } - onChanged(); - } - } else { - if (!other.columnFamilies_.isEmpty()) { - if (columnFamiliesBuilder_.isEmpty()) { - columnFamiliesBuilder_.dispose(); - columnFamiliesBuilder_ = null; - columnFamilies_ = other.columnFamilies_; - bitField0_ = (bitField0_ & ~0x00000004); - columnFamiliesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getColumnFamiliesFieldBuilder() : null; - } else { - columnFamiliesBuilder_.addAllMessages(other.columnFamilies_); - } - } - } - if (configurationBuilder_ == null) { - if (!other.configuration_.isEmpty()) { - if (configuration_.isEmpty()) { - configuration_ = other.configuration_; - bitField0_ = (bitField0_ & ~0x00000008); - } else { - ensureConfigurationIsMutable(); - configuration_.addAll(other.configuration_); - } - onChanged(); - } - } else { - if (!other.configuration_.isEmpty()) { - if (configurationBuilder_.isEmpty()) { - configurationBuilder_.dispose(); - configurationBuilder_ = null; - configuration_ = other.configuration_; - bitField0_ = (bitField0_ & ~0x00000008); - configurationBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getConfigurationFieldBuilder() : null; - } else { - configurationBuilder_.addAllMessages(other.configuration_); - } - } - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - for (int i = 0; i < getAttributesCount(); i++) { - if (!getAttributes(i).isInitialized()) { - - return false; - } - } - for (int i = 0; i < getColumnFamiliesCount(); i++) { - if (!getColumnFamilies(i).isInitialized()) { - - return false; - } - } - for (int i = 0; i < getConfigurationCount(); i++) { - if (!getConfiguration(i).isInitialized()) { - - return false; - } - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - name_ = input.readBytes(); - break; - } - case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addAttributes(subBuilder.buildPartial()); - break; - } - case 26: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addColumnFamilies(subBuilder.buildPartial()); - break; - } - case 34: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addConfiguration(subBuilder.buildPartial()); - break; - } - } - } - } - - private int bitField0_; - - // optional bytes name = 1; - private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public com.google.protobuf.ByteString getName() { - return name_; - } - public Builder setName(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - return this; - } - public Builder clearName() { - bitField0_ = (bitField0_ & ~0x00000001); - name_ = getDefaultInstance().getName(); - onChanged(); - return this; - } - - // repeated .BytesBytesPair attributes = 2; - private java.util.List attributes_ = - java.util.Collections.emptyList(); - private void ensureAttributesIsMutable() { - if (!((bitField0_ & 0x00000002) == 0x00000002)) { - attributes_ = new java.util.ArrayList(attributes_); - bitField0_ |= 0x00000002; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> attributesBuilder_; - - public java.util.List getAttributesList() { - if (attributesBuilder_ == null) { - return java.util.Collections.unmodifiableList(attributes_); - } else { - return attributesBuilder_.getMessageList(); - } - } - public int getAttributesCount() { - if (attributesBuilder_ == null) { - return attributes_.size(); - } else { - return attributesBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { - if (attributesBuilder_ == null) { - return attributes_.get(index); - } else { - return attributesBuilder_.getMessage(index); - } - } - public Builder setAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { - if (attributesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureAttributesIsMutable(); - attributes_.set(index, value); - onChanged(); - } else { - attributesBuilder_.setMessage(index, value); - } - return this; - } - public Builder setAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.set(index, builderForValue.build()); - onChanged(); - } else { - attributesBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAttributes(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { - if (attributesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureAttributesIsMutable(); - attributes_.add(value); - onChanged(); - } else { - attributesBuilder_.addMessage(value); - } - return this; - } - public Builder addAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { - if (attributesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureAttributesIsMutable(); - attributes_.add(index, value); - onChanged(); - } else { - attributesBuilder_.addMessage(index, value); - } - return this; - } - public Builder addAttributes( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.add(builderForValue.build()); - onChanged(); - } else { - attributesBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.add(index, builderForValue.build()); - onChanged(); - } else { - attributesBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllAttributes( - java.lang.Iterable values) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - super.addAll(values, attributes_); - onChanged(); - } else { - attributesBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearAttributes() { - if (attributesBuilder_ == null) { - attributes_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000002); - onChanged(); - } else { - attributesBuilder_.clear(); - } - return this; - } - public Builder removeAttributes(int index) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.remove(index); - onChanged(); - } else { - attributesBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder getAttributesBuilder( - int index) { - return getAttributesFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( - int index) { - if (attributesBuilder_ == null) { - return attributes_.get(index); } else { - return attributesBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getAttributesOrBuilderList() { - if (attributesBuilder_ != null) { - return attributesBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(attributes_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder() { - return getAttributesFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder( - int index) { - return getAttributesFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); - } - public java.util.List - getAttributesBuilderList() { - return getAttributesFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> - getAttributesFieldBuilder() { - if (attributesBuilder_ == null) { - attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>( - attributes_, - ((bitField0_ & 0x00000002) == 0x00000002), - getParentForChildren(), - isClean()); - attributes_ = null; - } - return attributesBuilder_; - } - - // repeated .ColumnFamilySchema columnFamilies = 3; - private java.util.List columnFamilies_ = - java.util.Collections.emptyList(); - private void ensureColumnFamiliesIsMutable() { - if (!((bitField0_ & 0x00000004) == 0x00000004)) { - columnFamilies_ = new java.util.ArrayList(columnFamilies_); - bitField0_ |= 0x00000004; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; - - public java.util.List getColumnFamiliesList() { - if (columnFamiliesBuilder_ == null) { - return java.util.Collections.unmodifiableList(columnFamilies_); - } else { - return columnFamiliesBuilder_.getMessageList(); - } - } - public int getColumnFamiliesCount() { - if (columnFamiliesBuilder_ == null) { - return columnFamilies_.size(); - } else { - return columnFamiliesBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index) { - if (columnFamiliesBuilder_ == null) { - return columnFamilies_.get(index); - } else { - return columnFamiliesBuilder_.getMessage(index); - } - } - public Builder setColumnFamilies( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { - if (columnFamiliesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureColumnFamiliesIsMutable(); - columnFamilies_.set(index, value); - onChanged(); - } else { - columnFamiliesBuilder_.setMessage(index, value); - } - return this; - } - public Builder setColumnFamilies( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { - if (columnFamiliesBuilder_ == null) { - ensureColumnFamiliesIsMutable(); - columnFamilies_.set(index, builderForValue.build()); - onChanged(); - } else { - columnFamiliesBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { - if (columnFamiliesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureColumnFamiliesIsMutable(); - columnFamilies_.add(value); - onChanged(); - } else { - columnFamiliesBuilder_.addMessage(value); - } - return this; - } - public Builder addColumnFamilies( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { - if (columnFamiliesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureColumnFamiliesIsMutable(); - columnFamilies_.add(index, value); - onChanged(); - } else { - columnFamiliesBuilder_.addMessage(index, value); - } - return this; - } - public Builder addColumnFamilies( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { - if (columnFamiliesBuilder_ == null) { - ensureColumnFamiliesIsMutable(); - columnFamilies_.add(builderForValue.build()); - onChanged(); - } else { - columnFamiliesBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addColumnFamilies( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { - if (columnFamiliesBuilder_ == null) { - ensureColumnFamiliesIsMutable(); - columnFamilies_.add(index, builderForValue.build()); - onChanged(); - } else { - columnFamiliesBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllColumnFamilies( - java.lang.Iterable values) { - if (columnFamiliesBuilder_ == null) { - ensureColumnFamiliesIsMutable(); - super.addAll(values, columnFamilies_); - onChanged(); - } else { - columnFamiliesBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearColumnFamilies() { - if (columnFamiliesBuilder_ == null) { - columnFamilies_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000004); - onChanged(); - } else { - columnFamiliesBuilder_.clear(); - } - return this; - } - public Builder removeColumnFamilies(int index) { - if (columnFamiliesBuilder_ == null) { - ensureColumnFamiliesIsMutable(); - columnFamilies_.remove(index); - onChanged(); - } else { - columnFamiliesBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder getColumnFamiliesBuilder( - int index) { - return getColumnFamiliesFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder( - int index) { - if (columnFamiliesBuilder_ == null) { - return columnFamilies_.get(index); } else { - return columnFamiliesBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getColumnFamiliesOrBuilderList() { - if (columnFamiliesBuilder_ != null) { - return columnFamiliesBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(columnFamilies_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder addColumnFamiliesBuilder() { - return getColumnFamiliesFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder addColumnFamiliesBuilder( - int index) { - return getColumnFamiliesFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()); - } - public java.util.List - getColumnFamiliesBuilderList() { - return getColumnFamiliesFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> - getColumnFamiliesFieldBuilder() { - if (columnFamiliesBuilder_ == null) { - columnFamiliesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>( - columnFamilies_, - ((bitField0_ & 0x00000004) == 0x00000004), - getParentForChildren(), - isClean()); - columnFamilies_ = null; - } - return columnFamiliesBuilder_; - } - - // repeated .NameStringPair configuration = 4; - private java.util.List configuration_ = - java.util.Collections.emptyList(); - private void ensureConfigurationIsMutable() { - if (!((bitField0_ & 0x00000008) == 0x00000008)) { - configuration_ = new java.util.ArrayList(configuration_); - bitField0_ |= 0x00000008; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_; - - public java.util.List getConfigurationList() { - if (configurationBuilder_ == null) { - return java.util.Collections.unmodifiableList(configuration_); - } else { - return configurationBuilder_.getMessageList(); - } - } - public int getConfigurationCount() { - if (configurationBuilder_ == null) { - return configuration_.size(); - } else { - return configurationBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { - if (configurationBuilder_ == null) { - return configuration_.get(index); - } else { - return configurationBuilder_.getMessage(index); - } - } - public Builder setConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { - if (configurationBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureConfigurationIsMutable(); - configuration_.set(index, value); - onChanged(); - } else { - configurationBuilder_.setMessage(index, value); - } - return this; - } - public Builder setConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.set(index, builderForValue.build()); - onChanged(); - } else { - configurationBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addConfiguration(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { - if (configurationBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureConfigurationIsMutable(); - configuration_.add(value); - onChanged(); - } else { - configurationBuilder_.addMessage(value); - } - return this; - } - public Builder addConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { - if (configurationBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureConfigurationIsMutable(); - configuration_.add(index, value); - onChanged(); - } else { - configurationBuilder_.addMessage(index, value); - } - return this; - } - public Builder addConfiguration( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.add(builderForValue.build()); - onChanged(); - } else { - configurationBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.add(index, builderForValue.build()); - onChanged(); - } else { - configurationBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllConfiguration( - java.lang.Iterable values) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - super.addAll(values, configuration_); - onChanged(); - } else { - configurationBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearConfiguration() { - if (configurationBuilder_ == null) { - configuration_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000008); - onChanged(); - } else { - configurationBuilder_.clear(); - } - return this; - } - public Builder removeConfiguration(int index) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.remove(index); - onChanged(); - } else { - configurationBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder getConfigurationBuilder( - int index) { - return getConfigurationFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( - int index) { - if (configurationBuilder_ == null) { - return configuration_.get(index); } else { - return configurationBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getConfigurationOrBuilderList() { - if (configurationBuilder_ != null) { - return configurationBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(configuration_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder() { - return getConfigurationFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder( - int index) { - return getConfigurationFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); - } - public java.util.List - getConfigurationBuilderList() { - return getConfigurationFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> - getConfigurationFieldBuilder() { - if (configurationBuilder_ == null) { - configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>( - configuration_, - ((bitField0_ & 0x00000008) == 0x00000008), - getParentForChildren(), - isClean()); - configuration_ = null; - } - return configurationBuilder_; - } - - // @@protoc_insertion_point(builder_scope:TableSchema) - } - - static { - defaultInstance = new TableSchema(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:TableSchema) - } - - public interface ColumnFamilySchemaOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required bytes name = 1; - boolean hasName(); - com.google.protobuf.ByteString getName(); - - // repeated .BytesBytesPair attributes = 2; - java.util.List - getAttributesList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index); - int getAttributesCount(); - java.util.List - getAttributesOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( - int index); - - // repeated .NameStringPair configuration = 3; - java.util.List - getConfigurationList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index); - int getConfigurationCount(); - java.util.List - getConfigurationOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( - int index); - } - public static final class ColumnFamilySchema extends - com.google.protobuf.GeneratedMessage - implements ColumnFamilySchemaOrBuilder { - // Use ColumnFamilySchema.newBuilder() to construct. - private ColumnFamilySchema(Builder builder) { - super(builder); - } - private ColumnFamilySchema(boolean noInit) {} - - private static final ColumnFamilySchema defaultInstance; - public static ColumnFamilySchema getDefaultInstance() { - return defaultInstance; - } - - public ColumnFamilySchema getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_fieldAccessorTable; - } - - private int bitField0_; - // required bytes name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString name_; - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public com.google.protobuf.ByteString getName() { - return name_; - } - - // repeated .BytesBytesPair attributes = 2; - public static final int ATTRIBUTES_FIELD_NUMBER = 2; - private java.util.List attributes_; - public java.util.List getAttributesList() { - return attributes_; - } - public java.util.List - getAttributesOrBuilderList() { - return attributes_; - } - public int getAttributesCount() { - return attributes_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { - return attributes_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( - int index) { - return attributes_.get(index); - } - - // repeated .NameStringPair configuration = 3; - public static final int CONFIGURATION_FIELD_NUMBER = 3; - private java.util.List configuration_; - public java.util.List getConfigurationList() { - return configuration_; - } - public java.util.List - getConfigurationOrBuilderList() { - return configuration_; - } - public int getConfigurationCount() { - return configuration_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { - return configuration_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( - int index) { - return configuration_.get(index); - } - - private void initFields() { - name_ = com.google.protobuf.ByteString.EMPTY; - attributes_ = java.util.Collections.emptyList(); - configuration_ = java.util.Collections.emptyList(); - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasName()) { - memoizedIsInitialized = 0; - return false; - } - for (int i = 0; i < getAttributesCount(); i++) { - if (!getAttributes(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - } - for (int i = 0; i < getConfigurationCount(); i++) { - if (!getConfiguration(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, name_); - } - for (int i = 0; i < attributes_.size(); i++) { - output.writeMessage(2, attributes_.get(i)); - } - for (int i = 0; i < configuration_.size(); i++) { - output.writeMessage(3, configuration_.get(i)); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, name_); - } - for (int i = 0; i < attributes_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, attributes_.get(i)); - } - for (int i = 0; i < configuration_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(3, configuration_.get(i)); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema) obj; - - boolean result = true; - result = result && (hasName() == other.hasName()); - if (hasName()) { - result = result && getName() - .equals(other.getName()); - } - result = result && getAttributesList() - .equals(other.getAttributesList()); - result = result && getConfigurationList() - .equals(other.getConfigurationList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - @java.lang.Override - public int hashCode() { - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasName()) { - hash = (37 * hash) + NAME_FIELD_NUMBER; - hash = (53 * hash) + getName().hashCode(); - } - if (getAttributesCount() > 0) { - hash = (37 * hash) + ATTRIBUTES_FIELD_NUMBER; - hash = (53 * hash) + getAttributesList().hashCode(); - } - if (getConfigurationCount() > 0) { - hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER; - hash = (53 * hash) + getConfigurationList().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_fieldAccessorTable; - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder(BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getAttributesFieldBuilder(); - getConfigurationFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - name_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000001); - if (attributesBuilder_ == null) { - attributes_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000002); - } else { - attributesBuilder_.clear(); - } - if (configurationBuilder_ == null) { - configuration_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000004); - } else { - configurationBuilder_.clear(); - } - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDescriptor(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.name_ = name_; - if (attributesBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002)) { - attributes_ = java.util.Collections.unmodifiableList(attributes_); - bitField0_ = (bitField0_ & ~0x00000002); - } - result.attributes_ = attributes_; - } else { - result.attributes_ = attributesBuilder_.build(); - } - if (configurationBuilder_ == null) { - if (((bitField0_ & 0x00000004) == 0x00000004)) { - configuration_ = java.util.Collections.unmodifiableList(configuration_); - bitField0_ = (bitField0_ & ~0x00000004); - } - result.configuration_ = configuration_; - } else { - result.configuration_ = configurationBuilder_.build(); - } - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()) return this; - if (other.hasName()) { - setName(other.getName()); - } - if (attributesBuilder_ == null) { - if (!other.attributes_.isEmpty()) { - if (attributes_.isEmpty()) { - attributes_ = other.attributes_; - bitField0_ = (bitField0_ & ~0x00000002); - } else { - ensureAttributesIsMutable(); - attributes_.addAll(other.attributes_); - } - onChanged(); - } - } else { - if (!other.attributes_.isEmpty()) { - if (attributesBuilder_.isEmpty()) { - attributesBuilder_.dispose(); - attributesBuilder_ = null; - attributes_ = other.attributes_; - bitField0_ = (bitField0_ & ~0x00000002); - attributesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getAttributesFieldBuilder() : null; - } else { - attributesBuilder_.addAllMessages(other.attributes_); - } - } - } - if (configurationBuilder_ == null) { - if (!other.configuration_.isEmpty()) { - if (configuration_.isEmpty()) { - configuration_ = other.configuration_; - bitField0_ = (bitField0_ & ~0x00000004); - } else { - ensureConfigurationIsMutable(); - configuration_.addAll(other.configuration_); - } - onChanged(); - } - } else { - if (!other.configuration_.isEmpty()) { - if (configurationBuilder_.isEmpty()) { - configurationBuilder_.dispose(); - configurationBuilder_ = null; - configuration_ = other.configuration_; - bitField0_ = (bitField0_ & ~0x00000004); - configurationBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getConfigurationFieldBuilder() : null; - } else { - configurationBuilder_.addAllMessages(other.configuration_); - } - } - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - if (!hasName()) { - - return false; - } - for (int i = 0; i < getAttributesCount(); i++) { - if (!getAttributes(i).isInitialized()) { - - return false; - } - } - for (int i = 0; i < getConfigurationCount(); i++) { - if (!getConfiguration(i).isInitialized()) { - - return false; - } - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - name_ = input.readBytes(); - break; - } - case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addAttributes(subBuilder.buildPartial()); - break; - } - case 26: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addConfiguration(subBuilder.buildPartial()); - break; - } - } - } - } - - private int bitField0_; - - // required bytes name = 1; - private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public com.google.protobuf.ByteString getName() { - return name_; - } - public Builder setName(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - return this; - } - public Builder clearName() { - bitField0_ = (bitField0_ & ~0x00000001); - name_ = getDefaultInstance().getName(); - onChanged(); - return this; - } - - // repeated .BytesBytesPair attributes = 2; - private java.util.List attributes_ = - java.util.Collections.emptyList(); - private void ensureAttributesIsMutable() { - if (!((bitField0_ & 0x00000002) == 0x00000002)) { - attributes_ = new java.util.ArrayList(attributes_); - bitField0_ |= 0x00000002; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> attributesBuilder_; - - public java.util.List getAttributesList() { - if (attributesBuilder_ == null) { - return java.util.Collections.unmodifiableList(attributes_); - } else { - return attributesBuilder_.getMessageList(); - } - } - public int getAttributesCount() { - if (attributesBuilder_ == null) { - return attributes_.size(); - } else { - return attributesBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { - if (attributesBuilder_ == null) { - return attributes_.get(index); - } else { - return attributesBuilder_.getMessage(index); - } - } - public Builder setAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { - if (attributesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureAttributesIsMutable(); - attributes_.set(index, value); - onChanged(); - } else { - attributesBuilder_.setMessage(index, value); - } - return this; - } - public Builder setAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.set(index, builderForValue.build()); - onChanged(); - } else { - attributesBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAttributes(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { - if (attributesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureAttributesIsMutable(); - attributes_.add(value); - onChanged(); - } else { - attributesBuilder_.addMessage(value); - } - return this; - } - public Builder addAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { - if (attributesBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureAttributesIsMutable(); - attributes_.add(index, value); - onChanged(); - } else { - attributesBuilder_.addMessage(index, value); - } - return this; - } - public Builder addAttributes( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.add(builderForValue.build()); - onChanged(); - } else { - attributesBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addAttributes( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.add(index, builderForValue.build()); - onChanged(); - } else { - attributesBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllAttributes( - java.lang.Iterable values) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - super.addAll(values, attributes_); - onChanged(); - } else { - attributesBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearAttributes() { - if (attributesBuilder_ == null) { - attributes_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000002); - onChanged(); - } else { - attributesBuilder_.clear(); - } - return this; - } - public Builder removeAttributes(int index) { - if (attributesBuilder_ == null) { - ensureAttributesIsMutable(); - attributes_.remove(index); - onChanged(); - } else { - attributesBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder getAttributesBuilder( - int index) { - return getAttributesFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( - int index) { - if (attributesBuilder_ == null) { - return attributes_.get(index); } else { - return attributesBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getAttributesOrBuilderList() { - if (attributesBuilder_ != null) { - return attributesBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(attributes_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder() { - return getAttributesFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder( - int index) { - return getAttributesFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); - } - public java.util.List - getAttributesBuilderList() { - return getAttributesFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> - getAttributesFieldBuilder() { - if (attributesBuilder_ == null) { - attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>( - attributes_, - ((bitField0_ & 0x00000002) == 0x00000002), - getParentForChildren(), - isClean()); - attributes_ = null; - } - return attributesBuilder_; - } - - // repeated .NameStringPair configuration = 3; - private java.util.List configuration_ = - java.util.Collections.emptyList(); - private void ensureConfigurationIsMutable() { - if (!((bitField0_ & 0x00000004) == 0x00000004)) { - configuration_ = new java.util.ArrayList(configuration_); - bitField0_ |= 0x00000004; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_; - - public java.util.List getConfigurationList() { - if (configurationBuilder_ == null) { - return java.util.Collections.unmodifiableList(configuration_); - } else { - return configurationBuilder_.getMessageList(); - } - } - public int getConfigurationCount() { - if (configurationBuilder_ == null) { - return configuration_.size(); - } else { - return configurationBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { - if (configurationBuilder_ == null) { - return configuration_.get(index); - } else { - return configurationBuilder_.getMessage(index); - } - } - public Builder setConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { - if (configurationBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureConfigurationIsMutable(); - configuration_.set(index, value); - onChanged(); - } else { - configurationBuilder_.setMessage(index, value); - } - return this; - } - public Builder setConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.set(index, builderForValue.build()); - onChanged(); - } else { - configurationBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addConfiguration(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { - if (configurationBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureConfigurationIsMutable(); - configuration_.add(value); - onChanged(); - } else { - configurationBuilder_.addMessage(value); - } - return this; - } - public Builder addConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { - if (configurationBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureConfigurationIsMutable(); - configuration_.add(index, value); - onChanged(); - } else { - configurationBuilder_.addMessage(index, value); - } - return this; - } - public Builder addConfiguration( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.add(builderForValue.build()); - onChanged(); - } else { - configurationBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addConfiguration( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.add(index, builderForValue.build()); - onChanged(); - } else { - configurationBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllConfiguration( - java.lang.Iterable values) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - super.addAll(values, configuration_); - onChanged(); - } else { - configurationBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearConfiguration() { - if (configurationBuilder_ == null) { - configuration_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000004); - onChanged(); - } else { - configurationBuilder_.clear(); - } - return this; - } - public Builder removeConfiguration(int index) { - if (configurationBuilder_ == null) { - ensureConfigurationIsMutable(); - configuration_.remove(index); - onChanged(); - } else { - configurationBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder getConfigurationBuilder( - int index) { - return getConfigurationFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( - int index) { - if (configurationBuilder_ == null) { - return configuration_.get(index); } else { - return configurationBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getConfigurationOrBuilderList() { - if (configurationBuilder_ != null) { - return configurationBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(configuration_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder() { - return getConfigurationFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder( - int index) { - return getConfigurationFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); - } - public java.util.List - getConfigurationBuilderList() { - return getConfigurationFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> - getConfigurationFieldBuilder() { - if (configurationBuilder_ == null) { - configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>( - configuration_, - ((bitField0_ & 0x00000004) == 0x00000004), - getParentForChildren(), - isClean()); - configuration_ = null; - } - return configurationBuilder_; - } - - // @@protoc_insertion_point(builder_scope:ColumnFamilySchema) - } - - static { - defaultInstance = new ColumnFamilySchema(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:ColumnFamilySchema) - } - - public interface RegionInfoOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required uint64 regionId = 1; - boolean hasRegionId(); - long getRegionId(); - - // required bytes tableName = 2; - boolean hasTableName(); - com.google.protobuf.ByteString getTableName(); - - // optional bytes startKey = 3; - boolean hasStartKey(); - com.google.protobuf.ByteString getStartKey(); - - // optional bytes endKey = 4; - boolean hasEndKey(); - com.google.protobuf.ByteString getEndKey(); - - // optional bool offline = 5; - boolean hasOffline(); - boolean getOffline(); - - // optional bool split = 6; - boolean hasSplit(); - boolean getSplit(); - - // optional bool recovering = 7; - boolean hasRecovering(); - boolean getRecovering(); - } - public static final class RegionInfo extends - com.google.protobuf.GeneratedMessage - implements RegionInfoOrBuilder { - // Use RegionInfo.newBuilder() to construct. - private RegionInfo(Builder builder) { - super(builder); - } - private RegionInfo(boolean noInit) {} - - private static final RegionInfo defaultInstance; - public static RegionInfo getDefaultInstance() { - return defaultInstance; - } - - public RegionInfo getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_fieldAccessorTable; - } - - private int bitField0_; - // required uint64 regionId = 1; - public static final int REGIONID_FIELD_NUMBER = 1; - private long regionId_; - public boolean hasRegionId() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public long getRegionId() { - return regionId_; - } - - // required bytes tableName = 2; - public static final int TABLENAME_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString tableName_; - public boolean hasTableName() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public com.google.protobuf.ByteString getTableName() { - return tableName_; - } - - // optional bytes startKey = 3; - public static final int STARTKEY_FIELD_NUMBER = 3; - private com.google.protobuf.ByteString startKey_; - public boolean hasStartKey() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public com.google.protobuf.ByteString getStartKey() { - return startKey_; - } - - // optional bytes endKey = 4; - public static final int ENDKEY_FIELD_NUMBER = 4; - private com.google.protobuf.ByteString endKey_; - public boolean hasEndKey() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public com.google.protobuf.ByteString getEndKey() { - return endKey_; - } - - // optional bool offline = 5; - public static final int OFFLINE_FIELD_NUMBER = 5; - private boolean offline_; - public boolean hasOffline() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - public boolean getOffline() { - return offline_; - } - - // optional bool split = 6; - public static final int SPLIT_FIELD_NUMBER = 6; - private boolean split_; - public boolean hasSplit() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - public boolean getSplit() { - return split_; - } - - // optional bool recovering = 7; - public static final int RECOVERING_FIELD_NUMBER = 7; - private boolean recovering_; - public boolean hasRecovering() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - public boolean getRecovering() { - return recovering_; - } - - private void initFields() { - regionId_ = 0L; - tableName_ = com.google.protobuf.ByteString.EMPTY; - startKey_ = com.google.protobuf.ByteString.EMPTY; - endKey_ = com.google.protobuf.ByteString.EMPTY; - offline_ = false; - split_ = false; - recovering_ = false; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasRegionId()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasTableName()) { - memoizedIsInitialized = 0; - return false; - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeUInt64(1, regionId_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, tableName_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, startKey_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeBytes(4, endKey_); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeBool(5, offline_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBool(6, split_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeBool(7, recovering_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(1, regionId_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, tableName_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, startKey_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(4, endKey_); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(5, offline_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(6, split_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(7, recovering_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo) obj; - - boolean result = true; - result = result && (hasRegionId() == other.hasRegionId()); - if (hasRegionId()) { - result = result && (getRegionId() - == other.getRegionId()); - } - result = result && (hasTableName() == other.hasTableName()); - if (hasTableName()) { - result = result && getTableName() - .equals(other.getTableName()); - } - result = result && (hasStartKey() == other.hasStartKey()); - if (hasStartKey()) { - result = result && getStartKey() - .equals(other.getStartKey()); - } - result = result && (hasEndKey() == other.hasEndKey()); - if (hasEndKey()) { - result = result && getEndKey() - .equals(other.getEndKey()); - } - result = result && (hasOffline() == other.hasOffline()); - if (hasOffline()) { - result = result && (getOffline() - == other.getOffline()); - } - result = result && (hasSplit() == other.hasSplit()); - if (hasSplit()) { - result = result && (getSplit() - == other.getSplit()); - } - result = result && (hasRecovering() == other.hasRecovering()); - if (hasRecovering()) { - result = result && (getRecovering() - == other.getRecovering()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - @java.lang.Override - public int hashCode() { - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasRegionId()) { - hash = (37 * hash) + REGIONID_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getRegionId()); - } - if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; - hash = (53 * hash) + getTableName().hashCode(); - } - if (hasStartKey()) { - hash = (37 * hash) + STARTKEY_FIELD_NUMBER; - hash = (53 * hash) + getStartKey().hashCode(); - } - if (hasEndKey()) { - hash = (37 * hash) + ENDKEY_FIELD_NUMBER; - hash = (53 * hash) + getEndKey().hashCode(); - } - if (hasOffline()) { - hash = (37 * hash) + OFFLINE_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getOffline()); - } - if (hasSplit()) { - hash = (37 * hash) + SPLIT_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getSplit()); - } - if (hasRecovering()) { - hash = (37 * hash) + RECOVERING_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getRecovering()); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_fieldAccessorTable; - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder(BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - regionId_ = 0L; - bitField0_ = (bitField0_ & ~0x00000001); - tableName_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000002); - startKey_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000004); - endKey_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000008); - offline_ = false; - bitField0_ = (bitField0_ & ~0x00000010); - split_ = false; - bitField0_ = (bitField0_ & ~0x00000020); - recovering_ = false; - bitField0_ = (bitField0_ & ~0x00000040); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDescriptor(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.regionId_ = regionId_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.tableName_ = tableName_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.startKey_ = startKey_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; - } - result.endKey_ = endKey_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; - } - result.offline_ = offline_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { - to_bitField0_ |= 0x00000020; - } - result.split_ = split_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { - to_bitField0_ |= 0x00000040; - } - result.recovering_ = recovering_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance()) return this; - if (other.hasRegionId()) { - setRegionId(other.getRegionId()); - } - if (other.hasTableName()) { - setTableName(other.getTableName()); - } - if (other.hasStartKey()) { - setStartKey(other.getStartKey()); - } - if (other.hasEndKey()) { - setEndKey(other.getEndKey()); - } - if (other.hasOffline()) { - setOffline(other.getOffline()); - } - if (other.hasSplit()) { - setSplit(other.getSplit()); - } - if (other.hasRecovering()) { - setRecovering(other.getRecovering()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - if (!hasRegionId()) { - - return false; - } - if (!hasTableName()) { - - return false; - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - } - break; - } - case 8: { - bitField0_ |= 0x00000001; - regionId_ = input.readUInt64(); - break; - } - case 18: { - bitField0_ |= 0x00000002; - tableName_ = input.readBytes(); - break; - } - case 26: { - bitField0_ |= 0x00000004; - startKey_ = input.readBytes(); - break; - } - case 34: { - bitField0_ |= 0x00000008; - endKey_ = input.readBytes(); - break; - } - case 40: { - bitField0_ |= 0x00000010; - offline_ = input.readBool(); - break; - } - case 48: { - bitField0_ |= 0x00000020; - split_ = input.readBool(); - break; - } - case 56: { - bitField0_ |= 0x00000040; - recovering_ = input.readBool(); - break; - } - } - } - } - - private int bitField0_; - - // required uint64 regionId = 1; - private long regionId_ ; - public boolean hasRegionId() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public long getRegionId() { - return regionId_; - } - public Builder setRegionId(long value) { - bitField0_ |= 0x00000001; - regionId_ = value; - onChanged(); - return this; - } - public Builder clearRegionId() { - bitField0_ = (bitField0_ & ~0x00000001); - regionId_ = 0L; - onChanged(); - return this; - } - - // required bytes tableName = 2; - private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasTableName() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public com.google.protobuf.ByteString getTableName() { - return tableName_; - } - public Builder setTableName(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - tableName_ = value; - onChanged(); - return this; - } - public Builder clearTableName() { - bitField0_ = (bitField0_ & ~0x00000002); - tableName_ = getDefaultInstance().getTableName(); - onChanged(); - return this; - } - - // optional bytes startKey = 3; - private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasStartKey() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public com.google.protobuf.ByteString getStartKey() { - return startKey_; - } - public Builder setStartKey(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000004; - startKey_ = value; - onChanged(); - return this; - } - public Builder clearStartKey() { - bitField0_ = (bitField0_ & ~0x00000004); - startKey_ = getDefaultInstance().getStartKey(); - onChanged(); - return this; - } - - // optional bytes endKey = 4; - private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasEndKey() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public com.google.protobuf.ByteString getEndKey() { - return endKey_; - } - public Builder setEndKey(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000008; - endKey_ = value; - onChanged(); - return this; - } - public Builder clearEndKey() { - bitField0_ = (bitField0_ & ~0x00000008); - endKey_ = getDefaultInstance().getEndKey(); - onChanged(); - return this; - } - - // optional bool offline = 5; - private boolean offline_ ; - public boolean hasOffline() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - public boolean getOffline() { - return offline_; - } - public Builder setOffline(boolean value) { - bitField0_ |= 0x00000010; - offline_ = value; - onChanged(); - return this; - } - public Builder clearOffline() { - bitField0_ = (bitField0_ & ~0x00000010); - offline_ = false; - onChanged(); - return this; - } - - // optional bool split = 6; - private boolean split_ ; - public boolean hasSplit() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - public boolean getSplit() { - return split_; - } - public Builder setSplit(boolean value) { - bitField0_ |= 0x00000020; - split_ = value; - onChanged(); - return this; - } - public Builder clearSplit() { - bitField0_ = (bitField0_ & ~0x00000020); - split_ = false; - onChanged(); - return this; - } - - // optional bool recovering = 7; - private boolean recovering_ ; - public boolean hasRecovering() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - public boolean getRecovering() { - return recovering_; - } - public Builder setRecovering(boolean value) { - bitField0_ |= 0x00000040; - recovering_ = value; - onChanged(); - return this; - } - public Builder clearRecovering() { - bitField0_ = (bitField0_ & ~0x00000040); - recovering_ = false; - onChanged(); - return this; - } - - // @@protoc_insertion_point(builder_scope:RegionInfo) - } - - static { - defaultInstance = new RegionInfo(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:RegionInfo) - } - - public interface FavoredNodesOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // repeated .ServerName favoredNode = 1; - java.util.List - getFavoredNodeList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index); - int getFavoredNodeCount(); - java.util.List - getFavoredNodeOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getFavoredNodeOrBuilder( - int index); - } - public static final class FavoredNodes extends - com.google.protobuf.GeneratedMessage - implements FavoredNodesOrBuilder { - // Use FavoredNodes.newBuilder() to construct. - private FavoredNodes(Builder builder) { - super(builder); - } - private FavoredNodes(boolean noInit) {} - - private static final FavoredNodes defaultInstance; - public static FavoredNodes getDefaultInstance() { - return defaultInstance; - } - - public FavoredNodes getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable; - } - - // repeated .ServerName favoredNode = 1; - public static final int FAVOREDNODE_FIELD_NUMBER = 1; - private java.util.List favoredNode_; - public java.util.List getFavoredNodeList() { - return favoredNode_; - } - public java.util.List - getFavoredNodeOrBuilderList() { - return favoredNode_; - } - public int getFavoredNodeCount() { - return favoredNode_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index) { - return favoredNode_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getFavoredNodeOrBuilder( - int index) { - return favoredNode_.get(index); - } - - private void initFields() { - favoredNode_ = java.util.Collections.emptyList(); - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - for (int i = 0; i < getFavoredNodeCount(); i++) { - if (!getFavoredNode(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - for (int i = 0; i < favoredNode_.size(); i++) { - output.writeMessage(1, favoredNode_.get(i)); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - for (int i = 0; i < favoredNode_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, favoredNode_.get(i)); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes) obj; - - boolean result = true; - result = result && getFavoredNodeList() - .equals(other.getFavoredNodeList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - @java.lang.Override - public int hashCode() { - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (getFavoredNodeCount() > 0) { - hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER; - hash = (53 * hash) + getFavoredNodeList().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodesOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable; - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder(BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getFavoredNodeFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - if (favoredNodeBuilder_ == null) { - favoredNode_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000001); - } else { - favoredNodeBuilder_.clear(); - } - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.getDescriptor(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes(this); - int from_bitField0_ = bitField0_; - if (favoredNodeBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001)) { - favoredNode_ = java.util.Collections.unmodifiableList(favoredNode_); - bitField0_ = (bitField0_ & ~0x00000001); - } - result.favoredNode_ = favoredNode_; - } else { - result.favoredNode_ = favoredNodeBuilder_.build(); - } - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.getDefaultInstance()) return this; - if (favoredNodeBuilder_ == null) { - if (!other.favoredNode_.isEmpty()) { - if (favoredNode_.isEmpty()) { - favoredNode_ = other.favoredNode_; - bitField0_ = (bitField0_ & ~0x00000001); - } else { - ensureFavoredNodeIsMutable(); - favoredNode_.addAll(other.favoredNode_); - } - onChanged(); - } - } else { - if (!other.favoredNode_.isEmpty()) { - if (favoredNodeBuilder_.isEmpty()) { - favoredNodeBuilder_.dispose(); - favoredNodeBuilder_ = null; - favoredNode_ = other.favoredNode_; - bitField0_ = (bitField0_ & ~0x00000001); - favoredNodeBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getFavoredNodeFieldBuilder() : null; - } else { - favoredNodeBuilder_.addAllMessages(other.favoredNode_); - } - } - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - for (int i = 0; i < getFavoredNodeCount(); i++) { - if (!getFavoredNode(i).isInitialized()) { - - return false; - } - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - } - break; - } - case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addFavoredNode(subBuilder.buildPartial()); - break; - } - } - } - } - - private int bitField0_; - - // repeated .ServerName favoredNode = 1; - private java.util.List favoredNode_ = - java.util.Collections.emptyList(); - private void ensureFavoredNodeIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { - favoredNode_ = new java.util.ArrayList(favoredNode_); - bitField0_ |= 0x00000001; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> favoredNodeBuilder_; - - public java.util.List getFavoredNodeList() { - if (favoredNodeBuilder_ == null) { - return java.util.Collections.unmodifiableList(favoredNode_); - } else { - return favoredNodeBuilder_.getMessageList(); - } - } - public int getFavoredNodeCount() { - if (favoredNodeBuilder_ == null) { - return favoredNode_.size(); - } else { - return favoredNodeBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index) { - if (favoredNodeBuilder_ == null) { - return favoredNode_.get(index); - } else { - return favoredNodeBuilder_.getMessage(index); - } - } - public Builder setFavoredNode( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { - if (favoredNodeBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureFavoredNodeIsMutable(); - favoredNode_.set(index, value); - onChanged(); - } else { - favoredNodeBuilder_.setMessage(index, value); - } - return this; - } - public Builder setFavoredNode( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { - if (favoredNodeBuilder_ == null) { - ensureFavoredNodeIsMutable(); - favoredNode_.set(index, builderForValue.build()); - onChanged(); - } else { - favoredNodeBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addFavoredNode(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { - if (favoredNodeBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureFavoredNodeIsMutable(); - favoredNode_.add(value); - onChanged(); - } else { - favoredNodeBuilder_.addMessage(value); - } - return this; - } - public Builder addFavoredNode( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { - if (favoredNodeBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureFavoredNodeIsMutable(); - favoredNode_.add(index, value); - onChanged(); - } else { - favoredNodeBuilder_.addMessage(index, value); - } - return this; - } - public Builder addFavoredNode( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { - if (favoredNodeBuilder_ == null) { - ensureFavoredNodeIsMutable(); - favoredNode_.add(builderForValue.build()); - onChanged(); - } else { - favoredNodeBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addFavoredNode( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { - if (favoredNodeBuilder_ == null) { - ensureFavoredNodeIsMutable(); - favoredNode_.add(index, builderForValue.build()); - onChanged(); - } else { - favoredNodeBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllFavoredNode( - java.lang.Iterable values) { - if (favoredNodeBuilder_ == null) { - ensureFavoredNodeIsMutable(); - super.addAll(values, favoredNode_); - onChanged(); - } else { - favoredNodeBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearFavoredNode() { - if (favoredNodeBuilder_ == null) { - favoredNode_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000001); - onChanged(); - } else { - favoredNodeBuilder_.clear(); - } - return this; - } - public Builder removeFavoredNode(int index) { - if (favoredNodeBuilder_ == null) { - ensureFavoredNodeIsMutable(); - favoredNode_.remove(index); - onChanged(); - } else { - favoredNodeBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getFavoredNodeBuilder( - int index) { - return getFavoredNodeFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getFavoredNodeOrBuilder( - int index) { - if (favoredNodeBuilder_ == null) { - return favoredNode_.get(index); } else { - return favoredNodeBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getFavoredNodeOrBuilderList() { - if (favoredNodeBuilder_ != null) { - return favoredNodeBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(favoredNode_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder addFavoredNodeBuilder() { - return getFavoredNodeFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder addFavoredNodeBuilder( - int index) { - return getFavoredNodeFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); - } - public java.util.List - getFavoredNodeBuilderList() { - return getFavoredNodeFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> - getFavoredNodeFieldBuilder() { - if (favoredNodeBuilder_ == null) { - favoredNodeBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( - favoredNode_, - ((bitField0_ & 0x00000001) == 0x00000001), - getParentForChildren(), - isClean()); - favoredNode_ = null; - } - return favoredNodeBuilder_; - } - - // @@protoc_insertion_point(builder_scope:FavoredNodes) - } - - static { - defaultInstance = new FavoredNodes(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:FavoredNodes) - } - - public interface RegionSpecifierOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required .RegionSpecifier.RegionSpecifierType type = 1; - boolean hasType(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType getType(); - - // required bytes value = 2; - boolean hasValue(); - com.google.protobuf.ByteString getValue(); - } - public static final class RegionSpecifier extends - com.google.protobuf.GeneratedMessage - implements RegionSpecifierOrBuilder { - // Use RegionSpecifier.newBuilder() to construct. - private RegionSpecifier(Builder builder) { - super(builder); - } - private RegionSpecifier(boolean noInit) {} - - private static final RegionSpecifier defaultInstance; - public static RegionSpecifier getDefaultInstance() { - return defaultInstance; - } - - public RegionSpecifier getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_fieldAccessorTable; - } - - public enum RegionSpecifierType - implements com.google.protobuf.ProtocolMessageEnum { - REGION_NAME(0, 1), - ENCODED_REGION_NAME(1, 2), - ; - - public static final int REGION_NAME_VALUE = 1; - public static final int ENCODED_REGION_NAME_VALUE = 2; - - - public final int getNumber() { return value; } - - public static RegionSpecifierType valueOf(int value) { - switch (value) { - case 1: return REGION_NAME; - case 2: return ENCODED_REGION_NAME; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap() { - public RegionSpecifierType findValueByNumber(int number) { - return RegionSpecifierType.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDescriptor().getEnumTypes().get(0); - } - - private static final RegionSpecifierType[] VALUES = { - REGION_NAME, ENCODED_REGION_NAME, - }; - - public static RegionSpecifierType valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private RegionSpecifierType(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:RegionSpecifier.RegionSpecifierType) - } - - private int bitField0_; - // required .RegionSpecifier.RegionSpecifierType type = 1; - public static final int TYPE_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType type_; - public boolean hasType() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType getType() { - return type_; - } - - // required bytes value = 2; - public static final int VALUE_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString value_; - public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public com.google.protobuf.ByteString getValue() { - return value_; - } - - private void initFields() { - type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; - value_ = com.google.protobuf.ByteString.EMPTY; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasType()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasValue()) { - memoizedIsInitialized = 0; - return false; - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeEnum(1, type_.getNumber()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, value_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeEnumSize(1, type_.getNumber()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, value_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier) obj; - - boolean result = true; - result = result && (hasType() == other.hasType()); - if (hasType()) { - result = result && - (getType() == other.getType()); - } - result = result && (hasValue() == other.hasValue()); - if (hasValue()) { - result = result && getValue() - .equals(other.getValue()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - @java.lang.Override - public int hashCode() { - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasType()) { - hash = (37 * hash) + TYPE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getType()); - } - if (hasValue()) { - hash = (37 * hash) + VALUE_FIELD_NUMBER; - hash = (53 * hash) + getValue().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_descriptor; + public Builder clearColumnFamilies() { + if (columnFamiliesBuilder_ == null) { + columnFamilies_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000004); + onChanged(); + } else { + columnFamiliesBuilder_.clear(); + } + return this; } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_fieldAccessorTable; + public Builder removeColumnFamilies(int index) { + if (columnFamiliesBuilder_ == null) { + ensureColumnFamiliesIsMutable(); + columnFamilies_.remove(index); + onChanged(); + } else { + columnFamiliesBuilder_.remove(index); + } + return this; } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder getColumnFamiliesBuilder( + int index) { + return getColumnFamiliesFieldBuilder().getBuilder(index); } - - private Builder(BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder( + int index) { + if (columnFamiliesBuilder_ == null) { + return columnFamilies_.get(index); } else { + return columnFamiliesBuilder_.getMessageOrBuilder(index); + } } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + public java.util.List + getColumnFamiliesOrBuilderList() { + if (columnFamiliesBuilder_ != null) { + return columnFamiliesBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(columnFamilies_); } } - private static Builder create() { - return new Builder(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder addColumnFamiliesBuilder() { + return getColumnFamiliesFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()); } - - public Builder clear() { - super.clear(); - type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; - bitField0_ = (bitField0_ & ~0x00000001); - value_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000002); - return this; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder addColumnFamiliesBuilder( + int index) { + return getColumnFamiliesFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()); } - - public Builder clone() { - return create().mergeFrom(buildPartial()); + public java.util.List + getColumnFamiliesBuilderList() { + return getColumnFamiliesFieldBuilder().getBuilderList(); } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDescriptor(); + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> + getColumnFamiliesFieldBuilder() { + if (columnFamiliesBuilder_ == null) { + columnFamiliesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>( + columnFamilies_, + ((bitField0_ & 0x00000004) == 0x00000004), + getParentForChildren(), + isClean()); + columnFamilies_ = null; + } + return columnFamiliesBuilder_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + // repeated .NameStringPair configuration = 4; + private java.util.List configuration_ = + java.util.Collections.emptyList(); + private void ensureConfigurationIsMutable() { + if (!((bitField0_ & 0x00000008) == 0x00000008)) { + configuration_ = new java.util.ArrayList(configuration_); + bitField0_ |= 0x00000008; + } } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); + public java.util.List getConfigurationList() { + if (configurationBuilder_ == null) { + return java.util.Collections.unmodifiableList(configuration_); + } else { + return configurationBuilder_.getMessageList(); } - return result; } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.type_ = type_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; + public int getConfigurationCount() { + if (configurationBuilder_ == null) { + return configuration_.size(); + } else { + return configurationBuilder_.getCount(); } - result.value_ = value_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier)other); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { + if (configurationBuilder_ == null) { + return configuration_.get(index); } else { - super.mergeFrom(other); - return this; + return configurationBuilder_.getMessage(index); } } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) return this; - if (other.hasType()) { - setType(other.getType()); - } - if (other.hasValue()) { - setValue(other.getValue()); + public Builder setConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { + if (configurationBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureConfigurationIsMutable(); + configuration_.set(index, value); + onChanged(); + } else { + configurationBuilder_.setMessage(index, value); } - this.mergeUnknownFields(other.getUnknownFields()); return this; } - - public final boolean isInitialized() { - if (!hasType()) { - - return false; + public Builder setConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.set(index, builderForValue.build()); + onChanged(); + } else { + configurationBuilder_.setMessage(index, builderForValue.build()); } - if (!hasValue()) { - - return false; + return this; + } + public Builder addConfiguration(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { + if (configurationBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureConfigurationIsMutable(); + configuration_.add(value); + onChanged(); + } else { + configurationBuilder_.addMessage(value); } - return true; + return this; } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - onChanged(); - return this; - } - break; - } - case 8: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType value = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(1, rawValue); - } else { - bitField0_ |= 0x00000001; - type_ = value; - } - break; - } - case 18: { - bitField0_ |= 0x00000002; - value_ = input.readBytes(); - break; - } + public Builder addConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { + if (configurationBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); } + ensureConfigurationIsMutable(); + configuration_.add(index, value); + onChanged(); + } else { + configurationBuilder_.addMessage(index, value); + } + return this; + } + public Builder addConfiguration( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.add(builderForValue.build()); + onChanged(); + } else { + configurationBuilder_.addMessage(builderForValue.build()); } + return this; } - - private int bitField0_; - - // required .RegionSpecifier.RegionSpecifierType type = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; - public boolean hasType() { - return ((bitField0_ & 0x00000001) == 0x00000001); + public Builder addConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.add(index, builderForValue.build()); + onChanged(); + } else { + configurationBuilder_.addMessage(index, builderForValue.build()); + } + return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType getType() { - return type_; + public Builder addAllConfiguration( + java.lang.Iterable values) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + super.addAll(values, configuration_); + onChanged(); + } else { + configurationBuilder_.addAllMessages(values); + } + return this; } - public Builder setType(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType value) { - if (value == null) { - throw new NullPointerException(); + public Builder clearConfiguration() { + if (configurationBuilder_ == null) { + configuration_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000008); + onChanged(); + } else { + configurationBuilder_.clear(); } - bitField0_ |= 0x00000001; - type_ = value; - onChanged(); return this; } - public Builder clearType() { - bitField0_ = (bitField0_ & ~0x00000001); - type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; - onChanged(); + public Builder removeConfiguration(int index) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.remove(index); + onChanged(); + } else { + configurationBuilder_.remove(index); + } return this; } - - // required bytes value = 2; - private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder getConfigurationBuilder( + int index) { + return getConfigurationFieldBuilder().getBuilder(index); } - public com.google.protobuf.ByteString getValue() { - return value_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( + int index) { + if (configurationBuilder_ == null) { + return configuration_.get(index); } else { + return configurationBuilder_.getMessageOrBuilder(index); + } } - public Builder setValue(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - value_ = value; - onChanged(); - return this; + public java.util.List + getConfigurationOrBuilderList() { + if (configurationBuilder_ != null) { + return configurationBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(configuration_); + } } - public Builder clearValue() { - bitField0_ = (bitField0_ & ~0x00000002); - value_ = getDefaultInstance().getValue(); - onChanged(); - return this; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder() { + return getConfigurationFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder( + int index) { + return getConfigurationFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); + } + public java.util.List + getConfigurationBuilderList() { + return getConfigurationFieldBuilder().getBuilderList(); + } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> + getConfigurationFieldBuilder() { + if (configurationBuilder_ == null) { + configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>( + configuration_, + ((bitField0_ & 0x00000008) == 0x00000008), + getParentForChildren(), + isClean()); + configuration_ = null; + } + return configurationBuilder_; } - // @@protoc_insertion_point(builder_scope:RegionSpecifier) + // @@protoc_insertion_point(builder_scope:TableSchema) } static { - defaultInstance = new RegionSpecifier(true); + defaultInstance = new TableSchema(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:RegionSpecifier) + // @@protoc_insertion_point(class_scope:TableSchema) } - public interface RegionLoadOrBuilder + public interface ColumnFamilySchemaOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .RegionSpecifier regionSpecifier = 1; - boolean hasRegionSpecifier(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder(); - - // optional uint32 stores = 2; - boolean hasStores(); - int getStores(); - - // optional uint32 storefiles = 3; - boolean hasStorefiles(); - int getStorefiles(); - - // optional uint32 storeUncompressedSizeMB = 4; - boolean hasStoreUncompressedSizeMB(); - int getStoreUncompressedSizeMB(); - - // optional uint32 storefileSizeMB = 5; - boolean hasStorefileSizeMB(); - int getStorefileSizeMB(); - - // optional uint32 memstoreSizeMB = 6; - boolean hasMemstoreSizeMB(); - int getMemstoreSizeMB(); - - // optional uint32 storefileIndexSizeMB = 7; - boolean hasStorefileIndexSizeMB(); - int getStorefileIndexSizeMB(); - - // optional uint64 readRequestsCount = 8; - boolean hasReadRequestsCount(); - long getReadRequestsCount(); - - // optional uint64 writeRequestsCount = 9; - boolean hasWriteRequestsCount(); - long getWriteRequestsCount(); - - // optional uint64 totalCompactingKVs = 10; - boolean hasTotalCompactingKVs(); - long getTotalCompactingKVs(); - - // optional uint64 currentCompactedKVs = 11; - boolean hasCurrentCompactedKVs(); - long getCurrentCompactedKVs(); - - // optional uint32 rootIndexSizeKB = 12; - boolean hasRootIndexSizeKB(); - int getRootIndexSizeKB(); - - // optional uint32 totalStaticIndexSizeKB = 13; - boolean hasTotalStaticIndexSizeKB(); - int getTotalStaticIndexSizeKB(); + // required bytes name = 1; + boolean hasName(); + com.google.protobuf.ByteString getName(); - // optional uint32 totalStaticBloomSizeKB = 14; - boolean hasTotalStaticBloomSizeKB(); - int getTotalStaticBloomSizeKB(); + // repeated .BytesBytesPair attributes = 2; + java.util.List + getAttributesList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index); + int getAttributesCount(); + java.util.List + getAttributesOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( + int index); - // optional uint64 completeSequenceId = 15; - boolean hasCompleteSequenceId(); - long getCompleteSequenceId(); + // repeated .NameStringPair configuration = 3; + java.util.List + getConfigurationList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index); + int getConfigurationCount(); + java.util.List + getConfigurationOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( + int index); } - public static final class RegionLoad extends + public static final class ColumnFamilySchema extends com.google.protobuf.GeneratedMessage - implements RegionLoadOrBuilder { - // Use RegionLoad.newBuilder() to construct. - private RegionLoad(Builder builder) { + implements ColumnFamilySchemaOrBuilder { + // Use ColumnFamilySchema.newBuilder() to construct. + private ColumnFamilySchema(Builder builder) { super(builder); } - private RegionLoad(boolean noInit) {} + private ColumnFamilySchema(boolean noInit) {} - private static final RegionLoad defaultInstance; - public static RegionLoad getDefaultInstance() { + private static final ColumnFamilySchema defaultInstance; + public static ColumnFamilySchema getDefaultInstance() { return defaultInstance; } - public RegionLoad getDefaultInstanceForType() { + public ColumnFamilySchema getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionLoad_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionLoad_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_fieldAccessorTable; } private int bitField0_; - // required .RegionSpecifier regionSpecifier = 1; - public static final int REGIONSPECIFIER_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_; - public boolean hasRegionSpecifier() { + // required bytes name = 1; + public static final int NAME_FIELD_NUMBER = 1; + private com.google.protobuf.ByteString name_; + public boolean hasName() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier() { - return regionSpecifier_; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder() { - return regionSpecifier_; - } - - // optional uint32 stores = 2; - public static final int STORES_FIELD_NUMBER = 2; - private int stores_; - public boolean hasStores() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public int getStores() { - return stores_; - } - - // optional uint32 storefiles = 3; - public static final int STOREFILES_FIELD_NUMBER = 3; - private int storefiles_; - public boolean hasStorefiles() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public int getStorefiles() { - return storefiles_; - } - - // optional uint32 storeUncompressedSizeMB = 4; - public static final int STOREUNCOMPRESSEDSIZEMB_FIELD_NUMBER = 4; - private int storeUncompressedSizeMB_; - public boolean hasStoreUncompressedSizeMB() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public int getStoreUncompressedSizeMB() { - return storeUncompressedSizeMB_; - } - - // optional uint32 storefileSizeMB = 5; - public static final int STOREFILESIZEMB_FIELD_NUMBER = 5; - private int storefileSizeMB_; - public boolean hasStorefileSizeMB() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - public int getStorefileSizeMB() { - return storefileSizeMB_; - } - - // optional uint32 memstoreSizeMB = 6; - public static final int MEMSTORESIZEMB_FIELD_NUMBER = 6; - private int memstoreSizeMB_; - public boolean hasMemstoreSizeMB() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - public int getMemstoreSizeMB() { - return memstoreSizeMB_; - } - - // optional uint32 storefileIndexSizeMB = 7; - public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 7; - private int storefileIndexSizeMB_; - public boolean hasStorefileIndexSizeMB() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - public int getStorefileIndexSizeMB() { - return storefileIndexSizeMB_; - } - - // optional uint64 readRequestsCount = 8; - public static final int READREQUESTSCOUNT_FIELD_NUMBER = 8; - private long readRequestsCount_; - public boolean hasReadRequestsCount() { - return ((bitField0_ & 0x00000080) == 0x00000080); - } - public long getReadRequestsCount() { - return readRequestsCount_; - } - - // optional uint64 writeRequestsCount = 9; - public static final int WRITEREQUESTSCOUNT_FIELD_NUMBER = 9; - private long writeRequestsCount_; - public boolean hasWriteRequestsCount() { - return ((bitField0_ & 0x00000100) == 0x00000100); - } - public long getWriteRequestsCount() { - return writeRequestsCount_; + public com.google.protobuf.ByteString getName() { + return name_; } - // optional uint64 totalCompactingKVs = 10; - public static final int TOTALCOMPACTINGKVS_FIELD_NUMBER = 10; - private long totalCompactingKVs_; - public boolean hasTotalCompactingKVs() { - return ((bitField0_ & 0x00000200) == 0x00000200); - } - public long getTotalCompactingKVs() { - return totalCompactingKVs_; + // repeated .BytesBytesPair attributes = 2; + public static final int ATTRIBUTES_FIELD_NUMBER = 2; + private java.util.List attributes_; + public java.util.List getAttributesList() { + return attributes_; } - - // optional uint64 currentCompactedKVs = 11; - public static final int CURRENTCOMPACTEDKVS_FIELD_NUMBER = 11; - private long currentCompactedKVs_; - public boolean hasCurrentCompactedKVs() { - return ((bitField0_ & 0x00000400) == 0x00000400); + public java.util.List + getAttributesOrBuilderList() { + return attributes_; } - public long getCurrentCompactedKVs() { - return currentCompactedKVs_; + public int getAttributesCount() { + return attributes_.size(); } - - // optional uint32 rootIndexSizeKB = 12; - public static final int ROOTINDEXSIZEKB_FIELD_NUMBER = 12; - private int rootIndexSizeKB_; - public boolean hasRootIndexSizeKB() { - return ((bitField0_ & 0x00000800) == 0x00000800); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { + return attributes_.get(index); } - public int getRootIndexSizeKB() { - return rootIndexSizeKB_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( + int index) { + return attributes_.get(index); } - // optional uint32 totalStaticIndexSizeKB = 13; - public static final int TOTALSTATICINDEXSIZEKB_FIELD_NUMBER = 13; - private int totalStaticIndexSizeKB_; - public boolean hasTotalStaticIndexSizeKB() { - return ((bitField0_ & 0x00001000) == 0x00001000); - } - public int getTotalStaticIndexSizeKB() { - return totalStaticIndexSizeKB_; + // repeated .NameStringPair configuration = 3; + public static final int CONFIGURATION_FIELD_NUMBER = 3; + private java.util.List configuration_; + public java.util.List getConfigurationList() { + return configuration_; } - - // optional uint32 totalStaticBloomSizeKB = 14; - public static final int TOTALSTATICBLOOMSIZEKB_FIELD_NUMBER = 14; - private int totalStaticBloomSizeKB_; - public boolean hasTotalStaticBloomSizeKB() { - return ((bitField0_ & 0x00002000) == 0x00002000); + public java.util.List + getConfigurationOrBuilderList() { + return configuration_; } - public int getTotalStaticBloomSizeKB() { - return totalStaticBloomSizeKB_; + public int getConfigurationCount() { + return configuration_.size(); } - - // optional uint64 completeSequenceId = 15; - public static final int COMPLETESEQUENCEID_FIELD_NUMBER = 15; - private long completeSequenceId_; - public boolean hasCompleteSequenceId() { - return ((bitField0_ & 0x00004000) == 0x00004000); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { + return configuration_.get(index); } - public long getCompleteSequenceId() { - return completeSequenceId_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( + int index) { + return configuration_.get(index); } private void initFields() { - regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); - stores_ = 0; - storefiles_ = 0; - storeUncompressedSizeMB_ = 0; - storefileSizeMB_ = 0; - memstoreSizeMB_ = 0; - storefileIndexSizeMB_ = 0; - readRequestsCount_ = 0L; - writeRequestsCount_ = 0L; - totalCompactingKVs_ = 0L; - currentCompactedKVs_ = 0L; - rootIndexSizeKB_ = 0; - totalStaticIndexSizeKB_ = 0; - totalStaticBloomSizeKB_ = 0; - completeSequenceId_ = 0L; + name_ = com.google.protobuf.ByteString.EMPTY; + attributes_ = java.util.Collections.emptyList(); + configuration_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - if (!hasRegionSpecifier()) { + if (!hasName()) { memoizedIsInitialized = 0; return false; } - if (!getRegionSpecifier().isInitialized()) { - memoizedIsInitialized = 0; - return false; + for (int i = 0; i < getAttributesCount(); i++) { + if (!getAttributes(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } + for (int i = 0; i < getConfigurationCount(); i++) { + if (!getConfiguration(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } } memoizedIsInitialized = 1; return true; @@ -5309,49 +1481,13 @@ public final class HBaseProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, regionSpecifier_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeUInt32(2, stores_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeUInt32(3, storefiles_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeUInt32(4, storeUncompressedSizeMB_); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeUInt32(5, storefileSizeMB_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeUInt32(6, memstoreSizeMB_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeUInt32(7, storefileIndexSizeMB_); - } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - output.writeUInt64(8, readRequestsCount_); - } - if (((bitField0_ & 0x00000100) == 0x00000100)) { - output.writeUInt64(9, writeRequestsCount_); - } - if (((bitField0_ & 0x00000200) == 0x00000200)) { - output.writeUInt64(10, totalCompactingKVs_); - } - if (((bitField0_ & 0x00000400) == 0x00000400)) { - output.writeUInt64(11, currentCompactedKVs_); - } - if (((bitField0_ & 0x00000800) == 0x00000800)) { - output.writeUInt32(12, rootIndexSizeKB_); - } - if (((bitField0_ & 0x00001000) == 0x00001000)) { - output.writeUInt32(13, totalStaticIndexSizeKB_); + output.writeBytes(1, name_); } - if (((bitField0_ & 0x00002000) == 0x00002000)) { - output.writeUInt32(14, totalStaticBloomSizeKB_); + for (int i = 0; i < attributes_.size(); i++) { + output.writeMessage(2, attributes_.get(i)); } - if (((bitField0_ & 0x00004000) == 0x00004000)) { - output.writeUInt64(15, completeSequenceId_); + for (int i = 0; i < configuration_.size(); i++) { + output.writeMessage(3, configuration_.get(i)); } getUnknownFields().writeTo(output); } @@ -5364,63 +1500,15 @@ public final class HBaseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, regionSpecifier_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(2, stores_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(3, storefiles_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(4, storeUncompressedSizeMB_); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(5, storefileSizeMB_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(6, memstoreSizeMB_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(7, storefileIndexSizeMB_); - } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(8, readRequestsCount_); - } - if (((bitField0_ & 0x00000100) == 0x00000100)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(9, writeRequestsCount_); - } - if (((bitField0_ & 0x00000200) == 0x00000200)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(10, totalCompactingKVs_); - } - if (((bitField0_ & 0x00000400) == 0x00000400)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(11, currentCompactedKVs_); - } - if (((bitField0_ & 0x00000800) == 0x00000800)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(12, rootIndexSizeKB_); - } - if (((bitField0_ & 0x00001000) == 0x00001000)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(13, totalStaticIndexSizeKB_); + .computeBytesSize(1, name_); } - if (((bitField0_ & 0x00002000) == 0x00002000)) { + for (int i = 0; i < attributes_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(14, totalStaticBloomSizeKB_); + .computeMessageSize(2, attributes_.get(i)); } - if (((bitField0_ & 0x00004000) == 0x00004000)) { + for (int i = 0; i < configuration_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(15, completeSequenceId_); + .computeMessageSize(3, configuration_.get(i)); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -5439,87 +1527,21 @@ public final class HBaseProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad) obj; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema) obj; boolean result = true; - result = result && (hasRegionSpecifier() == other.hasRegionSpecifier()); - if (hasRegionSpecifier()) { - result = result && getRegionSpecifier() - .equals(other.getRegionSpecifier()); - } - result = result && (hasStores() == other.hasStores()); - if (hasStores()) { - result = result && (getStores() - == other.getStores()); - } - result = result && (hasStorefiles() == other.hasStorefiles()); - if (hasStorefiles()) { - result = result && (getStorefiles() - == other.getStorefiles()); - } - result = result && (hasStoreUncompressedSizeMB() == other.hasStoreUncompressedSizeMB()); - if (hasStoreUncompressedSizeMB()) { - result = result && (getStoreUncompressedSizeMB() - == other.getStoreUncompressedSizeMB()); - } - result = result && (hasStorefileSizeMB() == other.hasStorefileSizeMB()); - if (hasStorefileSizeMB()) { - result = result && (getStorefileSizeMB() - == other.getStorefileSizeMB()); - } - result = result && (hasMemstoreSizeMB() == other.hasMemstoreSizeMB()); - if (hasMemstoreSizeMB()) { - result = result && (getMemstoreSizeMB() - == other.getMemstoreSizeMB()); - } - result = result && (hasStorefileIndexSizeMB() == other.hasStorefileIndexSizeMB()); - if (hasStorefileIndexSizeMB()) { - result = result && (getStorefileIndexSizeMB() - == other.getStorefileIndexSizeMB()); - } - result = result && (hasReadRequestsCount() == other.hasReadRequestsCount()); - if (hasReadRequestsCount()) { - result = result && (getReadRequestsCount() - == other.getReadRequestsCount()); - } - result = result && (hasWriteRequestsCount() == other.hasWriteRequestsCount()); - if (hasWriteRequestsCount()) { - result = result && (getWriteRequestsCount() - == other.getWriteRequestsCount()); - } - result = result && (hasTotalCompactingKVs() == other.hasTotalCompactingKVs()); - if (hasTotalCompactingKVs()) { - result = result && (getTotalCompactingKVs() - == other.getTotalCompactingKVs()); - } - result = result && (hasCurrentCompactedKVs() == other.hasCurrentCompactedKVs()); - if (hasCurrentCompactedKVs()) { - result = result && (getCurrentCompactedKVs() - == other.getCurrentCompactedKVs()); - } - result = result && (hasRootIndexSizeKB() == other.hasRootIndexSizeKB()); - if (hasRootIndexSizeKB()) { - result = result && (getRootIndexSizeKB() - == other.getRootIndexSizeKB()); - } - result = result && (hasTotalStaticIndexSizeKB() == other.hasTotalStaticIndexSizeKB()); - if (hasTotalStaticIndexSizeKB()) { - result = result && (getTotalStaticIndexSizeKB() - == other.getTotalStaticIndexSizeKB()); - } - result = result && (hasTotalStaticBloomSizeKB() == other.hasTotalStaticBloomSizeKB()); - if (hasTotalStaticBloomSizeKB()) { - result = result && (getTotalStaticBloomSizeKB() - == other.getTotalStaticBloomSizeKB()); - } - result = result && (hasCompleteSequenceId() == other.hasCompleteSequenceId()); - if (hasCompleteSequenceId()) { - result = result && (getCompleteSequenceId() - == other.getCompleteSequenceId()); + result = result && (hasName() == other.hasName()); + if (hasName()) { + result = result && getName() + .equals(other.getName()); } + result = result && getAttributesList() + .equals(other.getAttributesList()); + result = result && getConfigurationList() + .equals(other.getConfigurationList()); result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -5529,105 +1551,57 @@ public final class HBaseProtos { public int hashCode() { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasRegionSpecifier()) { - hash = (37 * hash) + REGIONSPECIFIER_FIELD_NUMBER; - hash = (53 * hash) + getRegionSpecifier().hashCode(); - } - if (hasStores()) { - hash = (37 * hash) + STORES_FIELD_NUMBER; - hash = (53 * hash) + getStores(); - } - if (hasStorefiles()) { - hash = (37 * hash) + STOREFILES_FIELD_NUMBER; - hash = (53 * hash) + getStorefiles(); - } - if (hasStoreUncompressedSizeMB()) { - hash = (37 * hash) + STOREUNCOMPRESSEDSIZEMB_FIELD_NUMBER; - hash = (53 * hash) + getStoreUncompressedSizeMB(); - } - if (hasStorefileSizeMB()) { - hash = (37 * hash) + STOREFILESIZEMB_FIELD_NUMBER; - hash = (53 * hash) + getStorefileSizeMB(); - } - if (hasMemstoreSizeMB()) { - hash = (37 * hash) + MEMSTORESIZEMB_FIELD_NUMBER; - hash = (53 * hash) + getMemstoreSizeMB(); - } - if (hasStorefileIndexSizeMB()) { - hash = (37 * hash) + STOREFILEINDEXSIZEMB_FIELD_NUMBER; - hash = (53 * hash) + getStorefileIndexSizeMB(); - } - if (hasReadRequestsCount()) { - hash = (37 * hash) + READREQUESTSCOUNT_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getReadRequestsCount()); - } - if (hasWriteRequestsCount()) { - hash = (37 * hash) + WRITEREQUESTSCOUNT_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getWriteRequestsCount()); - } - if (hasTotalCompactingKVs()) { - hash = (37 * hash) + TOTALCOMPACTINGKVS_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getTotalCompactingKVs()); - } - if (hasCurrentCompactedKVs()) { - hash = (37 * hash) + CURRENTCOMPACTEDKVS_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getCurrentCompactedKVs()); - } - if (hasRootIndexSizeKB()) { - hash = (37 * hash) + ROOTINDEXSIZEKB_FIELD_NUMBER; - hash = (53 * hash) + getRootIndexSizeKB(); - } - if (hasTotalStaticIndexSizeKB()) { - hash = (37 * hash) + TOTALSTATICINDEXSIZEKB_FIELD_NUMBER; - hash = (53 * hash) + getTotalStaticIndexSizeKB(); + if (hasName()) { + hash = (37 * hash) + NAME_FIELD_NUMBER; + hash = (53 * hash) + getName().hashCode(); } - if (hasTotalStaticBloomSizeKB()) { - hash = (37 * hash) + TOTALSTATICBLOOMSIZEKB_FIELD_NUMBER; - hash = (53 * hash) + getTotalStaticBloomSizeKB(); + if (getAttributesCount() > 0) { + hash = (37 * hash) + ATTRIBUTES_FIELD_NUMBER; + hash = (53 * hash) + getAttributesList().hashCode(); } - if (hasCompleteSequenceId()) { - hash = (37 * hash) + COMPLETESEQUENCEID_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getCompleteSequenceId()); + if (getConfigurationCount() > 0) { + hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER; + hash = (53 * hash) + getConfigurationList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -5636,7 +1610,7 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -5647,12 +1621,12 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -5662,7 +1636,7 @@ public final class HBaseProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -5675,18 +1649,18 @@ public final class HBaseProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionLoad_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionLoad_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ColumnFamilySchema_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -5697,7 +1671,8 @@ public final class HBaseProtos { } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getRegionSpecifierFieldBuilder(); + getAttributesFieldBuilder(); + getConfigurationFieldBuilder(); } } private static Builder create() { @@ -5706,40 +1681,20 @@ public final class HBaseProtos { public Builder clear() { super.clear(); - if (regionSpecifierBuilder_ == null) { - regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + name_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + if (attributesBuilder_ == null) { + attributes_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000002); } else { - regionSpecifierBuilder_.clear(); + attributesBuilder_.clear(); + } + if (configurationBuilder_ == null) { + configuration_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000004); + } else { + configurationBuilder_.clear(); } - bitField0_ = (bitField0_ & ~0x00000001); - stores_ = 0; - bitField0_ = (bitField0_ & ~0x00000002); - storefiles_ = 0; - bitField0_ = (bitField0_ & ~0x00000004); - storeUncompressedSizeMB_ = 0; - bitField0_ = (bitField0_ & ~0x00000008); - storefileSizeMB_ = 0; - bitField0_ = (bitField0_ & ~0x00000010); - memstoreSizeMB_ = 0; - bitField0_ = (bitField0_ & ~0x00000020); - storefileIndexSizeMB_ = 0; - bitField0_ = (bitField0_ & ~0x00000040); - readRequestsCount_ = 0L; - bitField0_ = (bitField0_ & ~0x00000080); - writeRequestsCount_ = 0L; - bitField0_ = (bitField0_ & ~0x00000100); - totalCompactingKVs_ = 0L; - bitField0_ = (bitField0_ & ~0x00000200); - currentCompactedKVs_ = 0L; - bitField0_ = (bitField0_ & ~0x00000400); - rootIndexSizeKB_ = 0; - bitField0_ = (bitField0_ & ~0x00000800); - totalStaticIndexSizeKB_ = 0; - bitField0_ = (bitField0_ & ~0x00001000); - totalStaticBloomSizeKB_ = 0; - bitField0_ = (bitField0_ & ~0x00002000); - completeSequenceId_ = 0L; - bitField0_ = (bitField0_ & ~0x00004000); return this; } @@ -5749,24 +1704,24 @@ public final class HBaseProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -5774,147 +1729,123 @@ public final class HBaseProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad(this); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - if (regionSpecifierBuilder_ == null) { - result.regionSpecifier_ = regionSpecifier_; + result.name_ = name_; + if (attributesBuilder_ == null) { + if (((bitField0_ & 0x00000002) == 0x00000002)) { + attributes_ = java.util.Collections.unmodifiableList(attributes_); + bitField0_ = (bitField0_ & ~0x00000002); + } + result.attributes_ = attributes_; } else { - result.regionSpecifier_ = regionSpecifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.stores_ = stores_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.storefiles_ = storefiles_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; - } - result.storeUncompressedSizeMB_ = storeUncompressedSizeMB_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; - } - result.storefileSizeMB_ = storefileSizeMB_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { - to_bitField0_ |= 0x00000020; - } - result.memstoreSizeMB_ = memstoreSizeMB_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { - to_bitField0_ |= 0x00000040; - } - result.storefileIndexSizeMB_ = storefileIndexSizeMB_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { - to_bitField0_ |= 0x00000080; - } - result.readRequestsCount_ = readRequestsCount_; - if (((from_bitField0_ & 0x00000100) == 0x00000100)) { - to_bitField0_ |= 0x00000100; - } - result.writeRequestsCount_ = writeRequestsCount_; - if (((from_bitField0_ & 0x00000200) == 0x00000200)) { - to_bitField0_ |= 0x00000200; - } - result.totalCompactingKVs_ = totalCompactingKVs_; - if (((from_bitField0_ & 0x00000400) == 0x00000400)) { - to_bitField0_ |= 0x00000400; - } - result.currentCompactedKVs_ = currentCompactedKVs_; - if (((from_bitField0_ & 0x00000800) == 0x00000800)) { - to_bitField0_ |= 0x00000800; - } - result.rootIndexSizeKB_ = rootIndexSizeKB_; - if (((from_bitField0_ & 0x00001000) == 0x00001000)) { - to_bitField0_ |= 0x00001000; - } - result.totalStaticIndexSizeKB_ = totalStaticIndexSizeKB_; - if (((from_bitField0_ & 0x00002000) == 0x00002000)) { - to_bitField0_ |= 0x00002000; + result.attributes_ = attributesBuilder_.build(); } - result.totalStaticBloomSizeKB_ = totalStaticBloomSizeKB_; - if (((from_bitField0_ & 0x00004000) == 0x00004000)) { - to_bitField0_ |= 0x00004000; + if (configurationBuilder_ == null) { + if (((bitField0_ & 0x00000004) == 0x00000004)) { + configuration_ = java.util.Collections.unmodifiableList(configuration_); + bitField0_ = (bitField0_ & ~0x00000004); + } + result.configuration_ = configuration_; + } else { + result.configuration_ = configurationBuilder_.build(); } - result.completeSequenceId_ = completeSequenceId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.getDefaultInstance()) return this; - if (other.hasRegionSpecifier()) { - mergeRegionSpecifier(other.getRegionSpecifier()); - } - if (other.hasStores()) { - setStores(other.getStores()); - } - if (other.hasStorefiles()) { - setStorefiles(other.getStorefiles()); - } - if (other.hasStoreUncompressedSizeMB()) { - setStoreUncompressedSizeMB(other.getStoreUncompressedSizeMB()); - } - if (other.hasStorefileSizeMB()) { - setStorefileSizeMB(other.getStorefileSizeMB()); - } - if (other.hasMemstoreSizeMB()) { - setMemstoreSizeMB(other.getMemstoreSizeMB()); - } - if (other.hasStorefileIndexSizeMB()) { - setStorefileIndexSizeMB(other.getStorefileIndexSizeMB()); - } - if (other.hasReadRequestsCount()) { - setReadRequestsCount(other.getReadRequestsCount()); - } - if (other.hasWriteRequestsCount()) { - setWriteRequestsCount(other.getWriteRequestsCount()); - } - if (other.hasTotalCompactingKVs()) { - setTotalCompactingKVs(other.getTotalCompactingKVs()); - } - if (other.hasCurrentCompactedKVs()) { - setCurrentCompactedKVs(other.getCurrentCompactedKVs()); - } - if (other.hasRootIndexSizeKB()) { - setRootIndexSizeKB(other.getRootIndexSizeKB()); - } - if (other.hasTotalStaticIndexSizeKB()) { - setTotalStaticIndexSizeKB(other.getTotalStaticIndexSizeKB()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()) return this; + if (other.hasName()) { + setName(other.getName()); } - if (other.hasTotalStaticBloomSizeKB()) { - setTotalStaticBloomSizeKB(other.getTotalStaticBloomSizeKB()); + if (attributesBuilder_ == null) { + if (!other.attributes_.isEmpty()) { + if (attributes_.isEmpty()) { + attributes_ = other.attributes_; + bitField0_ = (bitField0_ & ~0x00000002); + } else { + ensureAttributesIsMutable(); + attributes_.addAll(other.attributes_); + } + onChanged(); + } + } else { + if (!other.attributes_.isEmpty()) { + if (attributesBuilder_.isEmpty()) { + attributesBuilder_.dispose(); + attributesBuilder_ = null; + attributes_ = other.attributes_; + bitField0_ = (bitField0_ & ~0x00000002); + attributesBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getAttributesFieldBuilder() : null; + } else { + attributesBuilder_.addAllMessages(other.attributes_); + } + } } - if (other.hasCompleteSequenceId()) { - setCompleteSequenceId(other.getCompleteSequenceId()); + if (configurationBuilder_ == null) { + if (!other.configuration_.isEmpty()) { + if (configuration_.isEmpty()) { + configuration_ = other.configuration_; + bitField0_ = (bitField0_ & ~0x00000004); + } else { + ensureConfigurationIsMutable(); + configuration_.addAll(other.configuration_); + } + onChanged(); + } + } else { + if (!other.configuration_.isEmpty()) { + if (configurationBuilder_.isEmpty()) { + configurationBuilder_.dispose(); + configurationBuilder_ = null; + configuration_ = other.configuration_; + bitField0_ = (bitField0_ & ~0x00000004); + configurationBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getConfigurationFieldBuilder() : null; + } else { + configurationBuilder_.addAllMessages(other.configuration_); + } + } } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - if (!hasRegionSpecifier()) { + if (!hasName()) { return false; } - if (!getRegionSpecifier().isInitialized()) { - - return false; + for (int i = 0; i < getAttributesCount(); i++) { + if (!getAttributes(i).isInitialized()) { + + return false; + } + } + for (int i = 0; i < getConfigurationCount(); i++) { + if (!getConfiguration(i).isInitialized()) { + + return false; + } } return true; } @@ -5943,704 +1874,586 @@ public final class HBaseProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(); - if (hasRegionSpecifier()) { - subBuilder.mergeFrom(getRegionSpecifier()); - } - input.readMessage(subBuilder, extensionRegistry); - setRegionSpecifier(subBuilder.buildPartial()); - break; - } - case 16: { - bitField0_ |= 0x00000002; - stores_ = input.readUInt32(); - break; - } - case 24: { - bitField0_ |= 0x00000004; - storefiles_ = input.readUInt32(); - break; - } - case 32: { - bitField0_ |= 0x00000008; - storeUncompressedSizeMB_ = input.readUInt32(); - break; - } - case 40: { - bitField0_ |= 0x00000010; - storefileSizeMB_ = input.readUInt32(); - break; - } - case 48: { - bitField0_ |= 0x00000020; - memstoreSizeMB_ = input.readUInt32(); - break; - } - case 56: { - bitField0_ |= 0x00000040; - storefileIndexSizeMB_ = input.readUInt32(); - break; - } - case 64: { - bitField0_ |= 0x00000080; - readRequestsCount_ = input.readUInt64(); - break; - } - case 72: { - bitField0_ |= 0x00000100; - writeRequestsCount_ = input.readUInt64(); - break; - } - case 80: { - bitField0_ |= 0x00000200; - totalCompactingKVs_ = input.readUInt64(); - break; - } - case 88: { - bitField0_ |= 0x00000400; - currentCompactedKVs_ = input.readUInt64(); - break; - } - case 96: { - bitField0_ |= 0x00000800; - rootIndexSizeKB_ = input.readUInt32(); - break; - } - case 104: { - bitField0_ |= 0x00001000; - totalStaticIndexSizeKB_ = input.readUInt32(); + bitField0_ |= 0x00000001; + name_ = input.readBytes(); break; } - case 112: { - bitField0_ |= 0x00002000; - totalStaticBloomSizeKB_ = input.readUInt32(); + case 18: { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addAttributes(subBuilder.buildPartial()); break; } - case 120: { - bitField0_ |= 0x00004000; - completeSequenceId_ = input.readUInt64(); + case 26: { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addConfiguration(subBuilder.buildPartial()); break; } } } } - private int bitField0_; + private int bitField0_; + + // required bytes name = 1; + private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasName() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public com.google.protobuf.ByteString getName() { + return name_; + } + public Builder setName(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + name_ = value; + onChanged(); + return this; + } + public Builder clearName() { + bitField0_ = (bitField0_ & ~0x00000001); + name_ = getDefaultInstance().getName(); + onChanged(); + return this; + } + + // repeated .BytesBytesPair attributes = 2; + private java.util.List attributes_ = + java.util.Collections.emptyList(); + private void ensureAttributesIsMutable() { + if (!((bitField0_ & 0x00000002) == 0x00000002)) { + attributes_ = new java.util.ArrayList(attributes_); + bitField0_ |= 0x00000002; + } + } + + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> attributesBuilder_; + + public java.util.List getAttributesList() { + if (attributesBuilder_ == null) { + return java.util.Collections.unmodifiableList(attributes_); + } else { + return attributesBuilder_.getMessageList(); + } + } + public int getAttributesCount() { + if (attributesBuilder_ == null) { + return attributes_.size(); + } else { + return attributesBuilder_.getCount(); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getAttributes(int index) { + if (attributesBuilder_ == null) { + return attributes_.get(index); + } else { + return attributesBuilder_.getMessage(index); + } + } + public Builder setAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { + if (attributesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureAttributesIsMutable(); + attributes_.set(index, value); + onChanged(); + } else { + attributesBuilder_.setMessage(index, value); + } + return this; + } + public Builder setAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.set(index, builderForValue.build()); + onChanged(); + } else { + attributesBuilder_.setMessage(index, builderForValue.build()); + } + return this; + } + public Builder addAttributes(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { + if (attributesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureAttributesIsMutable(); + attributes_.add(value); + onChanged(); + } else { + attributesBuilder_.addMessage(value); + } + return this; + } + public Builder addAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { + if (attributesBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureAttributesIsMutable(); + attributes_.add(index, value); + onChanged(); + } else { + attributesBuilder_.addMessage(index, value); + } + return this; + } + public Builder addAttributes( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.add(builderForValue.build()); + onChanged(); + } else { + attributesBuilder_.addMessage(builderForValue.build()); + } + return this; + } + public Builder addAttributes( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.add(index, builderForValue.build()); + onChanged(); + } else { + attributesBuilder_.addMessage(index, builderForValue.build()); + } + return this; + } + public Builder addAllAttributes( + java.lang.Iterable values) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + super.addAll(values, attributes_); + onChanged(); + } else { + attributesBuilder_.addAllMessages(values); + } + return this; + } + public Builder clearAttributes() { + if (attributesBuilder_ == null) { + attributes_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000002); + onChanged(); + } else { + attributesBuilder_.clear(); + } + return this; + } + public Builder removeAttributes(int index) { + if (attributesBuilder_ == null) { + ensureAttributesIsMutable(); + attributes_.remove(index); + onChanged(); + } else { + attributesBuilder_.remove(index); + } + return this; + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder getAttributesBuilder( + int index) { + return getAttributesFieldBuilder().getBuilder(index); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( + int index) { + if (attributesBuilder_ == null) { + return attributes_.get(index); } else { + return attributesBuilder_.getMessageOrBuilder(index); + } + } + public java.util.List + getAttributesOrBuilderList() { + if (attributesBuilder_ != null) { + return attributesBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(attributes_); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder() { + return getAttributesFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addAttributesBuilder( + int index) { + return getAttributesFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); + } + public java.util.List + getAttributesBuilderList() { + return getAttributesFieldBuilder().getBuilderList(); + } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> + getAttributesFieldBuilder() { + if (attributesBuilder_ == null) { + attributesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>( + attributes_, + ((bitField0_ & 0x00000002) == 0x00000002), + getParentForChildren(), + isClean()); + attributes_ = null; + } + return attributesBuilder_; + } + + // repeated .NameStringPair configuration = 3; + private java.util.List configuration_ = + java.util.Collections.emptyList(); + private void ensureConfigurationIsMutable() { + if (!((bitField0_ & 0x00000004) == 0x00000004)) { + configuration_ = new java.util.ArrayList(configuration_); + bitField0_ |= 0x00000004; + } + } - // required .RegionSpecifier regionSpecifier = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionSpecifierBuilder_; - public boolean hasRegionSpecifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_; + + public java.util.List getConfigurationList() { + if (configurationBuilder_ == null) { + return java.util.Collections.unmodifiableList(configuration_); + } else { + return configurationBuilder_.getMessageList(); + } + } + public int getConfigurationCount() { + if (configurationBuilder_ == null) { + return configuration_.size(); + } else { + return configurationBuilder_.getCount(); + } } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier() { - if (regionSpecifierBuilder_ == null) { - return regionSpecifier_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { + if (configurationBuilder_ == null) { + return configuration_.get(index); } else { - return regionSpecifierBuilder_.getMessage(); + return configurationBuilder_.getMessage(index); } } - public Builder setRegionSpecifier(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionSpecifierBuilder_ == null) { + public Builder setConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { + if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } - regionSpecifier_ = value; + ensureConfigurationIsMutable(); + configuration_.set(index, value); onChanged(); } else { - regionSpecifierBuilder_.setMessage(value); + configurationBuilder_.setMessage(index, value); } - bitField0_ |= 0x00000001; return this; } - public Builder setRegionSpecifier( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { - if (regionSpecifierBuilder_ == null) { - regionSpecifier_ = builderForValue.build(); + public Builder setConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.set(index, builderForValue.build()); onChanged(); } else { - regionSpecifierBuilder_.setMessage(builderForValue.build()); + configurationBuilder_.setMessage(index, builderForValue.build()); } - bitField0_ |= 0x00000001; return this; } - public Builder mergeRegionSpecifier(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { - if (regionSpecifierBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && - regionSpecifier_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { - regionSpecifier_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionSpecifier_).mergeFrom(value).buildPartial(); - } else { - regionSpecifier_ = value; + public Builder addConfiguration(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { + if (configurationBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); } + ensureConfigurationIsMutable(); + configuration_.add(value); onChanged(); } else { - regionSpecifierBuilder_.mergeFrom(value); + configurationBuilder_.addMessage(value); } - bitField0_ |= 0x00000001; return this; } - public Builder clearRegionSpecifier() { - if (regionSpecifierBuilder_ == null) { - regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); + public Builder addConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { + if (configurationBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureConfigurationIsMutable(); + configuration_.add(index, value); onChanged(); } else { - regionSpecifierBuilder_.clear(); + configurationBuilder_.addMessage(index, value); } - bitField0_ = (bitField0_ & ~0x00000001); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionSpecifierBuilder() { - bitField0_ |= 0x00000001; - onChanged(); - return getRegionSpecifierFieldBuilder().getBuilder(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder() { - if (regionSpecifierBuilder_ != null) { - return regionSpecifierBuilder_.getMessageOrBuilder(); + public Builder addConfiguration( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.add(builderForValue.build()); + onChanged(); } else { - return regionSpecifier_; - } - } - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> - getRegionSpecifierFieldBuilder() { - if (regionSpecifierBuilder_ == null) { - regionSpecifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( - regionSpecifier_, - getParentForChildren(), - isClean()); - regionSpecifier_ = null; + configurationBuilder_.addMessage(builderForValue.build()); } - return regionSpecifierBuilder_; - } - - // optional uint32 stores = 2; - private int stores_ ; - public boolean hasStores() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public int getStores() { - return stores_; - } - public Builder setStores(int value) { - bitField0_ |= 0x00000002; - stores_ = value; - onChanged(); - return this; - } - public Builder clearStores() { - bitField0_ = (bitField0_ & ~0x00000002); - stores_ = 0; - onChanged(); - return this; - } - - // optional uint32 storefiles = 3; - private int storefiles_ ; - public boolean hasStorefiles() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public int getStorefiles() { - return storefiles_; - } - public Builder setStorefiles(int value) { - bitField0_ |= 0x00000004; - storefiles_ = value; - onChanged(); - return this; - } - public Builder clearStorefiles() { - bitField0_ = (bitField0_ & ~0x00000004); - storefiles_ = 0; - onChanged(); - return this; - } - - // optional uint32 storeUncompressedSizeMB = 4; - private int storeUncompressedSizeMB_ ; - public boolean hasStoreUncompressedSizeMB() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public int getStoreUncompressedSizeMB() { - return storeUncompressedSizeMB_; - } - public Builder setStoreUncompressedSizeMB(int value) { - bitField0_ |= 0x00000008; - storeUncompressedSizeMB_ = value; - onChanged(); - return this; - } - public Builder clearStoreUncompressedSizeMB() { - bitField0_ = (bitField0_ & ~0x00000008); - storeUncompressedSizeMB_ = 0; - onChanged(); - return this; - } - - // optional uint32 storefileSizeMB = 5; - private int storefileSizeMB_ ; - public boolean hasStorefileSizeMB() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - public int getStorefileSizeMB() { - return storefileSizeMB_; - } - public Builder setStorefileSizeMB(int value) { - bitField0_ |= 0x00000010; - storefileSizeMB_ = value; - onChanged(); - return this; - } - public Builder clearStorefileSizeMB() { - bitField0_ = (bitField0_ & ~0x00000010); - storefileSizeMB_ = 0; - onChanged(); - return this; - } - - // optional uint32 memstoreSizeMB = 6; - private int memstoreSizeMB_ ; - public boolean hasMemstoreSizeMB() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - public int getMemstoreSizeMB() { - return memstoreSizeMB_; - } - public Builder setMemstoreSizeMB(int value) { - bitField0_ |= 0x00000020; - memstoreSizeMB_ = value; - onChanged(); - return this; - } - public Builder clearMemstoreSizeMB() { - bitField0_ = (bitField0_ & ~0x00000020); - memstoreSizeMB_ = 0; - onChanged(); - return this; - } - - // optional uint32 storefileIndexSizeMB = 7; - private int storefileIndexSizeMB_ ; - public boolean hasStorefileIndexSizeMB() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - public int getStorefileIndexSizeMB() { - return storefileIndexSizeMB_; - } - public Builder setStorefileIndexSizeMB(int value) { - bitField0_ |= 0x00000040; - storefileIndexSizeMB_ = value; - onChanged(); - return this; - } - public Builder clearStorefileIndexSizeMB() { - bitField0_ = (bitField0_ & ~0x00000040); - storefileIndexSizeMB_ = 0; - onChanged(); - return this; - } - - // optional uint64 readRequestsCount = 8; - private long readRequestsCount_ ; - public boolean hasReadRequestsCount() { - return ((bitField0_ & 0x00000080) == 0x00000080); - } - public long getReadRequestsCount() { - return readRequestsCount_; - } - public Builder setReadRequestsCount(long value) { - bitField0_ |= 0x00000080; - readRequestsCount_ = value; - onChanged(); - return this; - } - public Builder clearReadRequestsCount() { - bitField0_ = (bitField0_ & ~0x00000080); - readRequestsCount_ = 0L; - onChanged(); - return this; - } - - // optional uint64 writeRequestsCount = 9; - private long writeRequestsCount_ ; - public boolean hasWriteRequestsCount() { - return ((bitField0_ & 0x00000100) == 0x00000100); - } - public long getWriteRequestsCount() { - return writeRequestsCount_; - } - public Builder setWriteRequestsCount(long value) { - bitField0_ |= 0x00000100; - writeRequestsCount_ = value; - onChanged(); - return this; - } - public Builder clearWriteRequestsCount() { - bitField0_ = (bitField0_ & ~0x00000100); - writeRequestsCount_ = 0L; - onChanged(); - return this; - } - - // optional uint64 totalCompactingKVs = 10; - private long totalCompactingKVs_ ; - public boolean hasTotalCompactingKVs() { - return ((bitField0_ & 0x00000200) == 0x00000200); - } - public long getTotalCompactingKVs() { - return totalCompactingKVs_; - } - public Builder setTotalCompactingKVs(long value) { - bitField0_ |= 0x00000200; - totalCompactingKVs_ = value; - onChanged(); - return this; - } - public Builder clearTotalCompactingKVs() { - bitField0_ = (bitField0_ & ~0x00000200); - totalCompactingKVs_ = 0L; - onChanged(); - return this; - } - - // optional uint64 currentCompactedKVs = 11; - private long currentCompactedKVs_ ; - public boolean hasCurrentCompactedKVs() { - return ((bitField0_ & 0x00000400) == 0x00000400); - } - public long getCurrentCompactedKVs() { - return currentCompactedKVs_; - } - public Builder setCurrentCompactedKVs(long value) { - bitField0_ |= 0x00000400; - currentCompactedKVs_ = value; - onChanged(); - return this; - } - public Builder clearCurrentCompactedKVs() { - bitField0_ = (bitField0_ & ~0x00000400); - currentCompactedKVs_ = 0L; - onChanged(); return this; } - - // optional uint32 rootIndexSizeKB = 12; - private int rootIndexSizeKB_ ; - public boolean hasRootIndexSizeKB() { - return ((bitField0_ & 0x00000800) == 0x00000800); - } - public int getRootIndexSizeKB() { - return rootIndexSizeKB_; - } - public Builder setRootIndexSizeKB(int value) { - bitField0_ |= 0x00000800; - rootIndexSizeKB_ = value; - onChanged(); + public Builder addConfiguration( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.add(index, builderForValue.build()); + onChanged(); + } else { + configurationBuilder_.addMessage(index, builderForValue.build()); + } return this; } - public Builder clearRootIndexSizeKB() { - bitField0_ = (bitField0_ & ~0x00000800); - rootIndexSizeKB_ = 0; - onChanged(); + public Builder addAllConfiguration( + java.lang.Iterable values) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + super.addAll(values, configuration_); + onChanged(); + } else { + configurationBuilder_.addAllMessages(values); + } return this; } - - // optional uint32 totalStaticIndexSizeKB = 13; - private int totalStaticIndexSizeKB_ ; - public boolean hasTotalStaticIndexSizeKB() { - return ((bitField0_ & 0x00001000) == 0x00001000); - } - public int getTotalStaticIndexSizeKB() { - return totalStaticIndexSizeKB_; - } - public Builder setTotalStaticIndexSizeKB(int value) { - bitField0_ |= 0x00001000; - totalStaticIndexSizeKB_ = value; - onChanged(); + public Builder clearConfiguration() { + if (configurationBuilder_ == null) { + configuration_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000004); + onChanged(); + } else { + configurationBuilder_.clear(); + } return this; } - public Builder clearTotalStaticIndexSizeKB() { - bitField0_ = (bitField0_ & ~0x00001000); - totalStaticIndexSizeKB_ = 0; - onChanged(); + public Builder removeConfiguration(int index) { + if (configurationBuilder_ == null) { + ensureConfigurationIsMutable(); + configuration_.remove(index); + onChanged(); + } else { + configurationBuilder_.remove(index); + } return this; } - - // optional uint32 totalStaticBloomSizeKB = 14; - private int totalStaticBloomSizeKB_ ; - public boolean hasTotalStaticBloomSizeKB() { - return ((bitField0_ & 0x00002000) == 0x00002000); - } - public int getTotalStaticBloomSizeKB() { - return totalStaticBloomSizeKB_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder getConfigurationBuilder( + int index) { + return getConfigurationFieldBuilder().getBuilder(index); } - public Builder setTotalStaticBloomSizeKB(int value) { - bitField0_ |= 0x00002000; - totalStaticBloomSizeKB_ = value; - onChanged(); - return this; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( + int index) { + if (configurationBuilder_ == null) { + return configuration_.get(index); } else { + return configurationBuilder_.getMessageOrBuilder(index); + } } - public Builder clearTotalStaticBloomSizeKB() { - bitField0_ = (bitField0_ & ~0x00002000); - totalStaticBloomSizeKB_ = 0; - onChanged(); - return this; + public java.util.List + getConfigurationOrBuilderList() { + if (configurationBuilder_ != null) { + return configurationBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(configuration_); + } } - - // optional uint64 completeSequenceId = 15; - private long completeSequenceId_ ; - public boolean hasCompleteSequenceId() { - return ((bitField0_ & 0x00004000) == 0x00004000); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder() { + return getConfigurationFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); } - public long getCompleteSequenceId() { - return completeSequenceId_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder( + int index) { + return getConfigurationFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); } - public Builder setCompleteSequenceId(long value) { - bitField0_ |= 0x00004000; - completeSequenceId_ = value; - onChanged(); - return this; + public java.util.List + getConfigurationBuilderList() { + return getConfigurationFieldBuilder().getBuilderList(); } - public Builder clearCompleteSequenceId() { - bitField0_ = (bitField0_ & ~0x00004000); - completeSequenceId_ = 0L; - onChanged(); - return this; + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> + getConfigurationFieldBuilder() { + if (configurationBuilder_ == null) { + configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>( + configuration_, + ((bitField0_ & 0x00000004) == 0x00000004), + getParentForChildren(), + isClean()); + configuration_ = null; + } + return configurationBuilder_; } - // @@protoc_insertion_point(builder_scope:RegionLoad) + // @@protoc_insertion_point(builder_scope:ColumnFamilySchema) } static { - defaultInstance = new RegionLoad(true); + defaultInstance = new ColumnFamilySchema(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:RegionLoad) + // @@protoc_insertion_point(class_scope:ColumnFamilySchema) } - public interface ServerLoadOrBuilder + public interface RegionInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint32 numberOfRequests = 1; - boolean hasNumberOfRequests(); - int getNumberOfRequests(); - - // optional uint32 totalNumberOfRequests = 2; - boolean hasTotalNumberOfRequests(); - int getTotalNumberOfRequests(); - - // optional uint32 usedHeapMB = 3; - boolean hasUsedHeapMB(); - int getUsedHeapMB(); - - // optional uint32 maxHeapMB = 4; - boolean hasMaxHeapMB(); - int getMaxHeapMB(); - - // repeated .RegionLoad regionLoads = 5; - java.util.List - getRegionLoadsList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad getRegionLoads(int index); - int getRegionLoadsCount(); - java.util.List - getRegionLoadsOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder getRegionLoadsOrBuilder( - int index); + // required uint64 regionId = 1; + boolean hasRegionId(); + long getRegionId(); - // repeated .Coprocessor coprocessors = 6; - java.util.List - getCoprocessorsList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getCoprocessors(int index); - int getCoprocessorsCount(); - java.util.List - getCoprocessorsOrBuilderList(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( - int index); + // required bytes tableName = 2; + boolean hasTableName(); + com.google.protobuf.ByteString getTableName(); + + // optional bytes startKey = 3; + boolean hasStartKey(); + com.google.protobuf.ByteString getStartKey(); + + // optional bytes endKey = 4; + boolean hasEndKey(); + com.google.protobuf.ByteString getEndKey(); - // optional uint64 reportStartTime = 7; - boolean hasReportStartTime(); - long getReportStartTime(); + // optional bool offline = 5; + boolean hasOffline(); + boolean getOffline(); - // optional uint64 reportEndTime = 8; - boolean hasReportEndTime(); - long getReportEndTime(); + // optional bool split = 6; + boolean hasSplit(); + boolean getSplit(); - // optional uint32 infoServerPort = 9; - boolean hasInfoServerPort(); - int getInfoServerPort(); + // optional bool recovering = 7; + boolean hasRecovering(); + boolean getRecovering(); } - public static final class ServerLoad extends + public static final class RegionInfo extends com.google.protobuf.GeneratedMessage - implements ServerLoadOrBuilder { - // Use ServerLoad.newBuilder() to construct. - private ServerLoad(Builder builder) { + implements RegionInfoOrBuilder { + // Use RegionInfo.newBuilder() to construct. + private RegionInfo(Builder builder) { super(builder); } - private ServerLoad(boolean noInit) {} + private RegionInfo(boolean noInit) {} - private static final ServerLoad defaultInstance; - public static ServerLoad getDefaultInstance() { + private static final RegionInfo defaultInstance; + public static RegionInfo getDefaultInstance() { return defaultInstance; } - public ServerLoad getDefaultInstanceForType() { + public RegionInfo getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ServerLoad_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ServerLoad_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_fieldAccessorTable; } private int bitField0_; - // optional uint32 numberOfRequests = 1; - public static final int NUMBEROFREQUESTS_FIELD_NUMBER = 1; - private int numberOfRequests_; - public boolean hasNumberOfRequests() { + // required uint64 regionId = 1; + public static final int REGIONID_FIELD_NUMBER = 1; + private long regionId_; + public boolean hasRegionId() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public int getNumberOfRequests() { - return numberOfRequests_; + public long getRegionId() { + return regionId_; } - // optional uint32 totalNumberOfRequests = 2; - public static final int TOTALNUMBEROFREQUESTS_FIELD_NUMBER = 2; - private int totalNumberOfRequests_; - public boolean hasTotalNumberOfRequests() { + // required bytes tableName = 2; + public static final int TABLENAME_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString tableName_; + public boolean hasTableName() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public int getTotalNumberOfRequests() { - return totalNumberOfRequests_; + public com.google.protobuf.ByteString getTableName() { + return tableName_; } - // optional uint32 usedHeapMB = 3; - public static final int USEDHEAPMB_FIELD_NUMBER = 3; - private int usedHeapMB_; - public boolean hasUsedHeapMB() { + // optional bytes startKey = 3; + public static final int STARTKEY_FIELD_NUMBER = 3; + private com.google.protobuf.ByteString startKey_; + public boolean hasStartKey() { return ((bitField0_ & 0x00000004) == 0x00000004); } - public int getUsedHeapMB() { - return usedHeapMB_; + public com.google.protobuf.ByteString getStartKey() { + return startKey_; } - // optional uint32 maxHeapMB = 4; - public static final int MAXHEAPMB_FIELD_NUMBER = 4; - private int maxHeapMB_; - public boolean hasMaxHeapMB() { + // optional bytes endKey = 4; + public static final int ENDKEY_FIELD_NUMBER = 4; + private com.google.protobuf.ByteString endKey_; + public boolean hasEndKey() { return ((bitField0_ & 0x00000008) == 0x00000008); } - public int getMaxHeapMB() { - return maxHeapMB_; - } - - // repeated .RegionLoad regionLoads = 5; - public static final int REGIONLOADS_FIELD_NUMBER = 5; - private java.util.List regionLoads_; - public java.util.List getRegionLoadsList() { - return regionLoads_; - } - public java.util.List - getRegionLoadsOrBuilderList() { - return regionLoads_; - } - public int getRegionLoadsCount() { - return regionLoads_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad getRegionLoads(int index) { - return regionLoads_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder getRegionLoadsOrBuilder( - int index) { - return regionLoads_.get(index); - } - - // repeated .Coprocessor coprocessors = 6; - public static final int COPROCESSORS_FIELD_NUMBER = 6; - private java.util.List coprocessors_; - public java.util.List getCoprocessorsList() { - return coprocessors_; - } - public java.util.List - getCoprocessorsOrBuilderList() { - return coprocessors_; - } - public int getCoprocessorsCount() { - return coprocessors_.size(); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getCoprocessors(int index) { - return coprocessors_.get(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( - int index) { - return coprocessors_.get(index); + public com.google.protobuf.ByteString getEndKey() { + return endKey_; } - // optional uint64 reportStartTime = 7; - public static final int REPORTSTARTTIME_FIELD_NUMBER = 7; - private long reportStartTime_; - public boolean hasReportStartTime() { + // optional bool offline = 5; + public static final int OFFLINE_FIELD_NUMBER = 5; + private boolean offline_; + public boolean hasOffline() { return ((bitField0_ & 0x00000010) == 0x00000010); } - public long getReportStartTime() { - return reportStartTime_; + public boolean getOffline() { + return offline_; } - // optional uint64 reportEndTime = 8; - public static final int REPORTENDTIME_FIELD_NUMBER = 8; - private long reportEndTime_; - public boolean hasReportEndTime() { + // optional bool split = 6; + public static final int SPLIT_FIELD_NUMBER = 6; + private boolean split_; + public boolean hasSplit() { return ((bitField0_ & 0x00000020) == 0x00000020); } - public long getReportEndTime() { - return reportEndTime_; + public boolean getSplit() { + return split_; } - // optional uint32 infoServerPort = 9; - public static final int INFOSERVERPORT_FIELD_NUMBER = 9; - private int infoServerPort_; - public boolean hasInfoServerPort() { + // optional bool recovering = 7; + public static final int RECOVERING_FIELD_NUMBER = 7; + private boolean recovering_; + public boolean hasRecovering() { return ((bitField0_ & 0x00000040) == 0x00000040); } - public int getInfoServerPort() { - return infoServerPort_; + public boolean getRecovering() { + return recovering_; } private void initFields() { - numberOfRequests_ = 0; - totalNumberOfRequests_ = 0; - usedHeapMB_ = 0; - maxHeapMB_ = 0; - regionLoads_ = java.util.Collections.emptyList(); - coprocessors_ = java.util.Collections.emptyList(); - reportStartTime_ = 0L; - reportEndTime_ = 0L; - infoServerPort_ = 0; + regionId_ = 0L; + tableName_ = com.google.protobuf.ByteString.EMPTY; + startKey_ = com.google.protobuf.ByteString.EMPTY; + endKey_ = com.google.protobuf.ByteString.EMPTY; + offline_ = false; + split_ = false; + recovering_ = false; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - for (int i = 0; i < getRegionLoadsCount(); i++) { - if (!getRegionLoads(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; - } + if (!hasRegionId()) { + memoizedIsInitialized = 0; + return false; } - for (int i = 0; i < getCoprocessorsCount(); i++) { - if (!getCoprocessors(i).isInitialized()) { - memoizedIsInitialized = 0; - return false; - } + if (!hasTableName()) { + memoizedIsInitialized = 0; + return false; } memoizedIsInitialized = 1; return true; @@ -6650,31 +2463,25 @@ public final class HBaseProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeUInt32(1, numberOfRequests_); + output.writeUInt64(1, regionId_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeUInt32(2, totalNumberOfRequests_); + output.writeBytes(2, tableName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeUInt32(3, usedHeapMB_); + output.writeBytes(3, startKey_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeUInt32(4, maxHeapMB_); - } - for (int i = 0; i < regionLoads_.size(); i++) { - output.writeMessage(5, regionLoads_.get(i)); - } - for (int i = 0; i < coprocessors_.size(); i++) { - output.writeMessage(6, coprocessors_.get(i)); + output.writeBytes(4, endKey_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeUInt64(7, reportStartTime_); + output.writeBool(5, offline_); } if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeUInt64(8, reportEndTime_); + output.writeBool(6, split_); } if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeUInt32(9, infoServerPort_); + output.writeBool(7, recovering_); } getUnknownFields().writeTo(output); } @@ -6687,39 +2494,31 @@ public final class HBaseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(1, numberOfRequests_); + .computeUInt64Size(1, regionId_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(2, totalNumberOfRequests_); + .computeBytesSize(2, tableName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(3, usedHeapMB_); + .computeBytesSize(3, startKey_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(4, maxHeapMB_); - } - for (int i = 0; i < regionLoads_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(5, regionLoads_.get(i)); - } - for (int i = 0; i < coprocessors_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(6, coprocessors_.get(i)); + .computeBytesSize(4, endKey_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(7, reportStartTime_); + .computeBoolSize(5, offline_); } if (((bitField0_ & 0x00000020) == 0x00000020)) { size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(8, reportEndTime_); + .computeBoolSize(6, split_); } if (((bitField0_ & 0x00000040) == 0x00000040)) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(9, infoServerPort_); + .computeBoolSize(7, recovering_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -6738,50 +2537,46 @@ public final class HBaseProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad) obj; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo) obj; boolean result = true; - result = result && (hasNumberOfRequests() == other.hasNumberOfRequests()); - if (hasNumberOfRequests()) { - result = result && (getNumberOfRequests() - == other.getNumberOfRequests()); - } - result = result && (hasTotalNumberOfRequests() == other.hasTotalNumberOfRequests()); - if (hasTotalNumberOfRequests()) { - result = result && (getTotalNumberOfRequests() - == other.getTotalNumberOfRequests()); - } - result = result && (hasUsedHeapMB() == other.hasUsedHeapMB()); - if (hasUsedHeapMB()) { - result = result && (getUsedHeapMB() - == other.getUsedHeapMB()); - } - result = result && (hasMaxHeapMB() == other.hasMaxHeapMB()); - if (hasMaxHeapMB()) { - result = result && (getMaxHeapMB() - == other.getMaxHeapMB()); - } - result = result && getRegionLoadsList() - .equals(other.getRegionLoadsList()); - result = result && getCoprocessorsList() - .equals(other.getCoprocessorsList()); - result = result && (hasReportStartTime() == other.hasReportStartTime()); - if (hasReportStartTime()) { - result = result && (getReportStartTime() - == other.getReportStartTime()); - } - result = result && (hasReportEndTime() == other.hasReportEndTime()); - if (hasReportEndTime()) { - result = result && (getReportEndTime() - == other.getReportEndTime()); - } - result = result && (hasInfoServerPort() == other.hasInfoServerPort()); - if (hasInfoServerPort()) { - result = result && (getInfoServerPort() - == other.getInfoServerPort()); + result = result && (hasRegionId() == other.hasRegionId()); + if (hasRegionId()) { + result = result && (getRegionId() + == other.getRegionId()); + } + result = result && (hasTableName() == other.hasTableName()); + if (hasTableName()) { + result = result && getTableName() + .equals(other.getTableName()); + } + result = result && (hasStartKey() == other.hasStartKey()); + if (hasStartKey()) { + result = result && getStartKey() + .equals(other.getStartKey()); + } + result = result && (hasEndKey() == other.hasEndKey()); + if (hasEndKey()) { + result = result && getEndKey() + .equals(other.getEndKey()); + } + result = result && (hasOffline() == other.hasOffline()); + if (hasOffline()) { + result = result && (getOffline() + == other.getOffline()); + } + result = result && (hasSplit() == other.hasSplit()); + if (hasSplit()) { + result = result && (getSplit() + == other.getSplit()); + } + result = result && (hasRecovering() == other.hasRecovering()); + if (hasRecovering()) { + result = result && (getRecovering() + == other.getRecovering()); } result = result && getUnknownFields().equals(other.getUnknownFields()); @@ -6792,81 +2587,73 @@ public final class HBaseProtos { public int hashCode() { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasNumberOfRequests()) { - hash = (37 * hash) + NUMBEROFREQUESTS_FIELD_NUMBER; - hash = (53 * hash) + getNumberOfRequests(); - } - if (hasTotalNumberOfRequests()) { - hash = (37 * hash) + TOTALNUMBEROFREQUESTS_FIELD_NUMBER; - hash = (53 * hash) + getTotalNumberOfRequests(); - } - if (hasUsedHeapMB()) { - hash = (37 * hash) + USEDHEAPMB_FIELD_NUMBER; - hash = (53 * hash) + getUsedHeapMB(); + if (hasRegionId()) { + hash = (37 * hash) + REGIONID_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getRegionId()); } - if (hasMaxHeapMB()) { - hash = (37 * hash) + MAXHEAPMB_FIELD_NUMBER; - hash = (53 * hash) + getMaxHeapMB(); + if (hasTableName()) { + hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (53 * hash) + getTableName().hashCode(); } - if (getRegionLoadsCount() > 0) { - hash = (37 * hash) + REGIONLOADS_FIELD_NUMBER; - hash = (53 * hash) + getRegionLoadsList().hashCode(); + if (hasStartKey()) { + hash = (37 * hash) + STARTKEY_FIELD_NUMBER; + hash = (53 * hash) + getStartKey().hashCode(); } - if (getCoprocessorsCount() > 0) { - hash = (37 * hash) + COPROCESSORS_FIELD_NUMBER; - hash = (53 * hash) + getCoprocessorsList().hashCode(); + if (hasEndKey()) { + hash = (37 * hash) + ENDKEY_FIELD_NUMBER; + hash = (53 * hash) + getEndKey().hashCode(); } - if (hasReportStartTime()) { - hash = (37 * hash) + REPORTSTARTTIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getReportStartTime()); + if (hasOffline()) { + hash = (37 * hash) + OFFLINE_FIELD_NUMBER; + hash = (53 * hash) + hashBoolean(getOffline()); } - if (hasReportEndTime()) { - hash = (37 * hash) + REPORTENDTIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getReportEndTime()); + if (hasSplit()) { + hash = (37 * hash) + SPLIT_FIELD_NUMBER; + hash = (53 * hash) + hashBoolean(getSplit()); } - if (hasInfoServerPort()) { - hash = (37 * hash) + INFOSERVERPORT_FIELD_NUMBER; - hash = (53 * hash) + getInfoServerPort(); + if (hasRecovering()) { + hash = (37 * hash) + RECOVERING_FIELD_NUMBER; + hash = (53 * hash) + hashBoolean(getRecovering()); } hash = (29 * hash) + getUnknownFields().hashCode(); return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -6875,7 +2662,7 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -6886,12 +2673,12 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -6901,7 +2688,7 @@ public final class HBaseProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -6914,18 +2701,18 @@ public final class HBaseProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ServerLoad_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_ServerLoad_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionInfo_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -6936,8 +2723,6 @@ public final class HBaseProtos { } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getRegionLoadsFieldBuilder(); - getCoprocessorsFieldBuilder(); } } private static Builder create() { @@ -6946,32 +2731,20 @@ public final class HBaseProtos { public Builder clear() { super.clear(); - numberOfRequests_ = 0; + regionId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); - totalNumberOfRequests_ = 0; + tableName_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); - usedHeapMB_ = 0; + startKey_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000004); - maxHeapMB_ = 0; + endKey_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000008); - if (regionLoadsBuilder_ == null) { - regionLoads_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000010); - } else { - regionLoadsBuilder_.clear(); - } - if (coprocessorsBuilder_ == null) { - coprocessors_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000020); - } else { - coprocessorsBuilder_.clear(); - } - reportStartTime_ = 0L; + offline_ = false; + bitField0_ = (bitField0_ & ~0x00000010); + split_ = false; + bitField0_ = (bitField0_ & ~0x00000020); + recovering_ = false; bitField0_ = (bitField0_ & ~0x00000040); - reportEndTime_ = 0L; - bitField0_ = (bitField0_ & ~0x00000080); - infoServerPort_ = 0; - bitField0_ = (bitField0_ & ~0x00000100); return this; } @@ -6981,24 +2754,24 @@ public final class HBaseProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -7006,161 +2779,87 @@ public final class HBaseProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad(this); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.numberOfRequests_ = numberOfRequests_; + result.regionId_ = regionId_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } - result.totalNumberOfRequests_ = totalNumberOfRequests_; + result.tableName_ = tableName_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } - result.usedHeapMB_ = usedHeapMB_; + result.startKey_ = startKey_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } - result.maxHeapMB_ = maxHeapMB_; - if (regionLoadsBuilder_ == null) { - if (((bitField0_ & 0x00000010) == 0x00000010)) { - regionLoads_ = java.util.Collections.unmodifiableList(regionLoads_); - bitField0_ = (bitField0_ & ~0x00000010); - } - result.regionLoads_ = regionLoads_; - } else { - result.regionLoads_ = regionLoadsBuilder_.build(); - } - if (coprocessorsBuilder_ == null) { - if (((bitField0_ & 0x00000020) == 0x00000020)) { - coprocessors_ = java.util.Collections.unmodifiableList(coprocessors_); - bitField0_ = (bitField0_ & ~0x00000020); - } - result.coprocessors_ = coprocessors_; - } else { - result.coprocessors_ = coprocessorsBuilder_.build(); - } - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + result.endKey_ = endKey_; + if (((from_bitField0_ & 0x00000010) == 0x00000010)) { to_bitField0_ |= 0x00000010; } - result.reportStartTime_ = reportStartTime_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + result.offline_ = offline_; + if (((from_bitField0_ & 0x00000020) == 0x00000020)) { to_bitField0_ |= 0x00000020; } - result.reportEndTime_ = reportEndTime_; - if (((from_bitField0_ & 0x00000100) == 0x00000100)) { + result.split_ = split_; + if (((from_bitField0_ & 0x00000040) == 0x00000040)) { to_bitField0_ |= 0x00000040; } - result.infoServerPort_ = infoServerPort_; + result.recovering_ = recovering_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance()) return this; - if (other.hasNumberOfRequests()) { - setNumberOfRequests(other.getNumberOfRequests()); - } - if (other.hasTotalNumberOfRequests()) { - setTotalNumberOfRequests(other.getTotalNumberOfRequests()); - } - if (other.hasUsedHeapMB()) { - setUsedHeapMB(other.getUsedHeapMB()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance()) return this; + if (other.hasRegionId()) { + setRegionId(other.getRegionId()); } - if (other.hasMaxHeapMB()) { - setMaxHeapMB(other.getMaxHeapMB()); + if (other.hasTableName()) { + setTableName(other.getTableName()); } - if (regionLoadsBuilder_ == null) { - if (!other.regionLoads_.isEmpty()) { - if (regionLoads_.isEmpty()) { - regionLoads_ = other.regionLoads_; - bitField0_ = (bitField0_ & ~0x00000010); - } else { - ensureRegionLoadsIsMutable(); - regionLoads_.addAll(other.regionLoads_); - } - onChanged(); - } - } else { - if (!other.regionLoads_.isEmpty()) { - if (regionLoadsBuilder_.isEmpty()) { - regionLoadsBuilder_.dispose(); - regionLoadsBuilder_ = null; - regionLoads_ = other.regionLoads_; - bitField0_ = (bitField0_ & ~0x00000010); - regionLoadsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getRegionLoadsFieldBuilder() : null; - } else { - regionLoadsBuilder_.addAllMessages(other.regionLoads_); - } - } + if (other.hasStartKey()) { + setStartKey(other.getStartKey()); } - if (coprocessorsBuilder_ == null) { - if (!other.coprocessors_.isEmpty()) { - if (coprocessors_.isEmpty()) { - coprocessors_ = other.coprocessors_; - bitField0_ = (bitField0_ & ~0x00000020); - } else { - ensureCoprocessorsIsMutable(); - coprocessors_.addAll(other.coprocessors_); - } - onChanged(); - } - } else { - if (!other.coprocessors_.isEmpty()) { - if (coprocessorsBuilder_.isEmpty()) { - coprocessorsBuilder_.dispose(); - coprocessorsBuilder_ = null; - coprocessors_ = other.coprocessors_; - bitField0_ = (bitField0_ & ~0x00000020); - coprocessorsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getCoprocessorsFieldBuilder() : null; - } else { - coprocessorsBuilder_.addAllMessages(other.coprocessors_); - } - } + if (other.hasEndKey()) { + setEndKey(other.getEndKey()); } - if (other.hasReportStartTime()) { - setReportStartTime(other.getReportStartTime()); + if (other.hasOffline()) { + setOffline(other.getOffline()); } - if (other.hasReportEndTime()) { - setReportEndTime(other.getReportEndTime()); + if (other.hasSplit()) { + setSplit(other.getSplit()); } - if (other.hasInfoServerPort()) { - setInfoServerPort(other.getInfoServerPort()); + if (other.hasRecovering()) { + setRecovering(other.getRecovering()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - for (int i = 0; i < getRegionLoadsCount(); i++) { - if (!getRegionLoads(i).isInitialized()) { - - return false; - } + if (!hasRegionId()) { + + return false; } - for (int i = 0; i < getCoprocessorsCount(); i++) { - if (!getCoprocessors(i).isInitialized()) { - - return false; - } + if (!hasTableName()) { + + return false; } return true; } @@ -7190,668 +2889,297 @@ public final class HBaseProtos { } case 8: { bitField0_ |= 0x00000001; - numberOfRequests_ = input.readUInt32(); + regionId_ = input.readUInt64(); break; } - case 16: { + case 18: { bitField0_ |= 0x00000002; - totalNumberOfRequests_ = input.readUInt32(); + tableName_ = input.readBytes(); break; } - case 24: { + case 26: { bitField0_ |= 0x00000004; - usedHeapMB_ = input.readUInt32(); + startKey_ = input.readBytes(); break; } - case 32: { + case 34: { bitField0_ |= 0x00000008; - maxHeapMB_ = input.readUInt32(); - break; - } - case 42: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addRegionLoads(subBuilder.buildPartial()); - break; - } - case 50: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addCoprocessors(subBuilder.buildPartial()); - break; - } - case 56: { - bitField0_ |= 0x00000040; - reportStartTime_ = input.readUInt64(); + endKey_ = input.readBytes(); break; } - case 64: { - bitField0_ |= 0x00000080; - reportEndTime_ = input.readUInt64(); + case 40: { + bitField0_ |= 0x00000010; + offline_ = input.readBool(); break; } - case 72: { - bitField0_ |= 0x00000100; - infoServerPort_ = input.readUInt32(); + case 48: { + bitField0_ |= 0x00000020; + split_ = input.readBool(); + break; + } + case 56: { + bitField0_ |= 0x00000040; + recovering_ = input.readBool(); break; } } } } - - private int bitField0_; - - // optional uint32 numberOfRequests = 1; - private int numberOfRequests_ ; - public boolean hasNumberOfRequests() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public int getNumberOfRequests() { - return numberOfRequests_; - } - public Builder setNumberOfRequests(int value) { - bitField0_ |= 0x00000001; - numberOfRequests_ = value; - onChanged(); - return this; - } - public Builder clearNumberOfRequests() { - bitField0_ = (bitField0_ & ~0x00000001); - numberOfRequests_ = 0; - onChanged(); - return this; - } - - // optional uint32 totalNumberOfRequests = 2; - private int totalNumberOfRequests_ ; - public boolean hasTotalNumberOfRequests() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - public int getTotalNumberOfRequests() { - return totalNumberOfRequests_; - } - public Builder setTotalNumberOfRequests(int value) { - bitField0_ |= 0x00000002; - totalNumberOfRequests_ = value; - onChanged(); - return this; - } - public Builder clearTotalNumberOfRequests() { - bitField0_ = (bitField0_ & ~0x00000002); - totalNumberOfRequests_ = 0; - onChanged(); - return this; - } - - // optional uint32 usedHeapMB = 3; - private int usedHeapMB_ ; - public boolean hasUsedHeapMB() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public int getUsedHeapMB() { - return usedHeapMB_; - } - public Builder setUsedHeapMB(int value) { - bitField0_ |= 0x00000004; - usedHeapMB_ = value; - onChanged(); - return this; - } - public Builder clearUsedHeapMB() { - bitField0_ = (bitField0_ & ~0x00000004); - usedHeapMB_ = 0; - onChanged(); - return this; - } - - // optional uint32 maxHeapMB = 4; - private int maxHeapMB_ ; - public boolean hasMaxHeapMB() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public int getMaxHeapMB() { - return maxHeapMB_; - } - public Builder setMaxHeapMB(int value) { - bitField0_ |= 0x00000008; - maxHeapMB_ = value; - onChanged(); - return this; - } - public Builder clearMaxHeapMB() { - bitField0_ = (bitField0_ & ~0x00000008); - maxHeapMB_ = 0; - onChanged(); - return this; - } - - // repeated .RegionLoad regionLoads = 5; - private java.util.List regionLoads_ = - java.util.Collections.emptyList(); - private void ensureRegionLoadsIsMutable() { - if (!((bitField0_ & 0x00000010) == 0x00000010)) { - regionLoads_ = new java.util.ArrayList(regionLoads_); - bitField0_ |= 0x00000010; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder> regionLoadsBuilder_; - - public java.util.List getRegionLoadsList() { - if (regionLoadsBuilder_ == null) { - return java.util.Collections.unmodifiableList(regionLoads_); - } else { - return regionLoadsBuilder_.getMessageList(); - } - } - public int getRegionLoadsCount() { - if (regionLoadsBuilder_ == null) { - return regionLoads_.size(); - } else { - return regionLoadsBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad getRegionLoads(int index) { - if (regionLoadsBuilder_ == null) { - return regionLoads_.get(index); - } else { - return regionLoadsBuilder_.getMessage(index); - } - } - public Builder setRegionLoads( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad value) { - if (regionLoadsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureRegionLoadsIsMutable(); - regionLoads_.set(index, value); - onChanged(); - } else { - regionLoadsBuilder_.setMessage(index, value); - } - return this; - } - public Builder setRegionLoads( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder builderForValue) { - if (regionLoadsBuilder_ == null) { - ensureRegionLoadsIsMutable(); - regionLoads_.set(index, builderForValue.build()); - onChanged(); - } else { - regionLoadsBuilder_.setMessage(index, builderForValue.build()); - } - return this; - } - public Builder addRegionLoads(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad value) { - if (regionLoadsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureRegionLoadsIsMutable(); - regionLoads_.add(value); - onChanged(); - } else { - regionLoadsBuilder_.addMessage(value); - } - return this; - } - public Builder addRegionLoads( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad value) { - if (regionLoadsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureRegionLoadsIsMutable(); - regionLoads_.add(index, value); - onChanged(); - } else { - regionLoadsBuilder_.addMessage(index, value); - } - return this; - } - public Builder addRegionLoads( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder builderForValue) { - if (regionLoadsBuilder_ == null) { - ensureRegionLoadsIsMutable(); - regionLoads_.add(builderForValue.build()); - onChanged(); - } else { - regionLoadsBuilder_.addMessage(builderForValue.build()); - } - return this; - } - public Builder addRegionLoads( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder builderForValue) { - if (regionLoadsBuilder_ == null) { - ensureRegionLoadsIsMutable(); - regionLoads_.add(index, builderForValue.build()); - onChanged(); - } else { - regionLoadsBuilder_.addMessage(index, builderForValue.build()); - } - return this; - } - public Builder addAllRegionLoads( - java.lang.Iterable values) { - if (regionLoadsBuilder_ == null) { - ensureRegionLoadsIsMutable(); - super.addAll(values, regionLoads_); - onChanged(); - } else { - regionLoadsBuilder_.addAllMessages(values); - } - return this; - } - public Builder clearRegionLoads() { - if (regionLoadsBuilder_ == null) { - regionLoads_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000010); - onChanged(); - } else { - regionLoadsBuilder_.clear(); - } - return this; - } - public Builder removeRegionLoads(int index) { - if (regionLoadsBuilder_ == null) { - ensureRegionLoadsIsMutable(); - regionLoads_.remove(index); - onChanged(); - } else { - regionLoadsBuilder_.remove(index); - } - return this; - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder getRegionLoadsBuilder( - int index) { - return getRegionLoadsFieldBuilder().getBuilder(index); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder getRegionLoadsOrBuilder( - int index) { - if (regionLoadsBuilder_ == null) { - return regionLoads_.get(index); } else { - return regionLoadsBuilder_.getMessageOrBuilder(index); - } - } - public java.util.List - getRegionLoadsOrBuilderList() { - if (regionLoadsBuilder_ != null) { - return regionLoadsBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(regionLoads_); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder addRegionLoadsBuilder() { - return getRegionLoadsFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.getDefaultInstance()); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder addRegionLoadsBuilder( - int index) { - return getRegionLoadsFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.getDefaultInstance()); - } - public java.util.List - getRegionLoadsBuilderList() { - return getRegionLoadsFieldBuilder().getBuilderList(); - } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder> - getRegionLoadsFieldBuilder() { - if (regionLoadsBuilder_ == null) { - regionLoadsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoadOrBuilder>( - regionLoads_, - ((bitField0_ & 0x00000010) == 0x00000010), - getParentForChildren(), - isClean()); - regionLoads_ = null; - } - return regionLoadsBuilder_; - } - - // repeated .Coprocessor coprocessors = 6; - private java.util.List coprocessors_ = - java.util.Collections.emptyList(); - private void ensureCoprocessorsIsMutable() { - if (!((bitField0_ & 0x00000020) == 0x00000020)) { - coprocessors_ = new java.util.ArrayList(coprocessors_); - bitField0_ |= 0x00000020; - } - } - - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder> coprocessorsBuilder_; - - public java.util.List getCoprocessorsList() { - if (coprocessorsBuilder_ == null) { - return java.util.Collections.unmodifiableList(coprocessors_); - } else { - return coprocessorsBuilder_.getMessageList(); - } - } - public int getCoprocessorsCount() { - if (coprocessorsBuilder_ == null) { - return coprocessors_.size(); - } else { - return coprocessorsBuilder_.getCount(); - } - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getCoprocessors(int index) { - if (coprocessorsBuilder_ == null) { - return coprocessors_.get(index); - } else { - return coprocessorsBuilder_.getMessage(index); - } - } - public Builder setCoprocessors( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor value) { - if (coprocessorsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureCoprocessorsIsMutable(); - coprocessors_.set(index, value); - onChanged(); - } else { - coprocessorsBuilder_.setMessage(index, value); - } - return this; - } - public Builder setCoprocessors( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder builderForValue) { - if (coprocessorsBuilder_ == null) { - ensureCoprocessorsIsMutable(); - coprocessors_.set(index, builderForValue.build()); - onChanged(); - } else { - coprocessorsBuilder_.setMessage(index, builderForValue.build()); - } - return this; + + private int bitField0_; + + // required uint64 regionId = 1; + private long regionId_ ; + public boolean hasRegionId() { + return ((bitField0_ & 0x00000001) == 0x00000001); } - public Builder addCoprocessors(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor value) { - if (coprocessorsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureCoprocessorsIsMutable(); - coprocessors_.add(value); - onChanged(); - } else { - coprocessorsBuilder_.addMessage(value); - } - return this; + public long getRegionId() { + return regionId_; } - public Builder addCoprocessors( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor value) { - if (coprocessorsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ensureCoprocessorsIsMutable(); - coprocessors_.add(index, value); - onChanged(); - } else { - coprocessorsBuilder_.addMessage(index, value); - } + public Builder setRegionId(long value) { + bitField0_ |= 0x00000001; + regionId_ = value; + onChanged(); return this; } - public Builder addCoprocessors( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder builderForValue) { - if (coprocessorsBuilder_ == null) { - ensureCoprocessorsIsMutable(); - coprocessors_.add(builderForValue.build()); - onChanged(); - } else { - coprocessorsBuilder_.addMessage(builderForValue.build()); - } + public Builder clearRegionId() { + bitField0_ = (bitField0_ & ~0x00000001); + regionId_ = 0L; + onChanged(); return this; } - public Builder addCoprocessors( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder builderForValue) { - if (coprocessorsBuilder_ == null) { - ensureCoprocessorsIsMutable(); - coprocessors_.add(index, builderForValue.build()); - onChanged(); - } else { - coprocessorsBuilder_.addMessage(index, builderForValue.build()); - } - return this; + + // required bytes tableName = 2; + private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasTableName() { + return ((bitField0_ & 0x00000002) == 0x00000002); } - public Builder addAllCoprocessors( - java.lang.Iterable values) { - if (coprocessorsBuilder_ == null) { - ensureCoprocessorsIsMutable(); - super.addAll(values, coprocessors_); - onChanged(); - } else { - coprocessorsBuilder_.addAllMessages(values); - } - return this; + public com.google.protobuf.ByteString getTableName() { + return tableName_; } - public Builder clearCoprocessors() { - if (coprocessorsBuilder_ == null) { - coprocessors_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000020); - onChanged(); - } else { - coprocessorsBuilder_.clear(); - } + public Builder setTableName(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + tableName_ = value; + onChanged(); return this; } - public Builder removeCoprocessors(int index) { - if (coprocessorsBuilder_ == null) { - ensureCoprocessorsIsMutable(); - coprocessors_.remove(index); - onChanged(); - } else { - coprocessorsBuilder_.remove(index); - } + public Builder clearTableName() { + bitField0_ = (bitField0_ & ~0x00000002); + tableName_ = getDefaultInstance().getTableName(); + onChanged(); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder getCoprocessorsBuilder( - int index) { - return getCoprocessorsFieldBuilder().getBuilder(index); + + // optional bytes startKey = 3; + private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasStartKey() { + return ((bitField0_ & 0x00000004) == 0x00000004); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( - int index) { - if (coprocessorsBuilder_ == null) { - return coprocessors_.get(index); } else { - return coprocessorsBuilder_.getMessageOrBuilder(index); - } + public com.google.protobuf.ByteString getStartKey() { + return startKey_; } - public java.util.List - getCoprocessorsOrBuilderList() { - if (coprocessorsBuilder_ != null) { - return coprocessorsBuilder_.getMessageOrBuilderList(); - } else { - return java.util.Collections.unmodifiableList(coprocessors_); - } + public Builder setStartKey(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000004; + startKey_ = value; + onChanged(); + return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder addCoprocessorsBuilder() { - return getCoprocessorsFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.getDefaultInstance()); + public Builder clearStartKey() { + bitField0_ = (bitField0_ & ~0x00000004); + startKey_ = getDefaultInstance().getStartKey(); + onChanged(); + return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder addCoprocessorsBuilder( - int index) { - return getCoprocessorsFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.getDefaultInstance()); + + // optional bytes endKey = 4; + private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasEndKey() { + return ((bitField0_ & 0x00000008) == 0x00000008); } - public java.util.List - getCoprocessorsBuilderList() { - return getCoprocessorsFieldBuilder().getBuilderList(); + public com.google.protobuf.ByteString getEndKey() { + return endKey_; } - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder> - getCoprocessorsFieldBuilder() { - if (coprocessorsBuilder_ == null) { - coprocessorsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder>( - coprocessors_, - ((bitField0_ & 0x00000020) == 0x00000020), - getParentForChildren(), - isClean()); - coprocessors_ = null; - } - return coprocessorsBuilder_; + public Builder setEndKey(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000008; + endKey_ = value; + onChanged(); + return this; + } + public Builder clearEndKey() { + bitField0_ = (bitField0_ & ~0x00000008); + endKey_ = getDefaultInstance().getEndKey(); + onChanged(); + return this; } - // optional uint64 reportStartTime = 7; - private long reportStartTime_ ; - public boolean hasReportStartTime() { - return ((bitField0_ & 0x00000040) == 0x00000040); + // optional bool offline = 5; + private boolean offline_ ; + public boolean hasOffline() { + return ((bitField0_ & 0x00000010) == 0x00000010); } - public long getReportStartTime() { - return reportStartTime_; + public boolean getOffline() { + return offline_; } - public Builder setReportStartTime(long value) { - bitField0_ |= 0x00000040; - reportStartTime_ = value; + public Builder setOffline(boolean value) { + bitField0_ |= 0x00000010; + offline_ = value; onChanged(); return this; } - public Builder clearReportStartTime() { - bitField0_ = (bitField0_ & ~0x00000040); - reportStartTime_ = 0L; + public Builder clearOffline() { + bitField0_ = (bitField0_ & ~0x00000010); + offline_ = false; onChanged(); return this; } - // optional uint64 reportEndTime = 8; - private long reportEndTime_ ; - public boolean hasReportEndTime() { - return ((bitField0_ & 0x00000080) == 0x00000080); + // optional bool split = 6; + private boolean split_ ; + public boolean hasSplit() { + return ((bitField0_ & 0x00000020) == 0x00000020); } - public long getReportEndTime() { - return reportEndTime_; + public boolean getSplit() { + return split_; } - public Builder setReportEndTime(long value) { - bitField0_ |= 0x00000080; - reportEndTime_ = value; + public Builder setSplit(boolean value) { + bitField0_ |= 0x00000020; + split_ = value; onChanged(); return this; } - public Builder clearReportEndTime() { - bitField0_ = (bitField0_ & ~0x00000080); - reportEndTime_ = 0L; + public Builder clearSplit() { + bitField0_ = (bitField0_ & ~0x00000020); + split_ = false; onChanged(); return this; } - // optional uint32 infoServerPort = 9; - private int infoServerPort_ ; - public boolean hasInfoServerPort() { - return ((bitField0_ & 0x00000100) == 0x00000100); + // optional bool recovering = 7; + private boolean recovering_ ; + public boolean hasRecovering() { + return ((bitField0_ & 0x00000040) == 0x00000040); } - public int getInfoServerPort() { - return infoServerPort_; + public boolean getRecovering() { + return recovering_; } - public Builder setInfoServerPort(int value) { - bitField0_ |= 0x00000100; - infoServerPort_ = value; + public Builder setRecovering(boolean value) { + bitField0_ |= 0x00000040; + recovering_ = value; onChanged(); return this; } - public Builder clearInfoServerPort() { - bitField0_ = (bitField0_ & ~0x00000100); - infoServerPort_ = 0; + public Builder clearRecovering() { + bitField0_ = (bitField0_ & ~0x00000040); + recovering_ = false; onChanged(); return this; } - // @@protoc_insertion_point(builder_scope:ServerLoad) + // @@protoc_insertion_point(builder_scope:RegionInfo) } static { - defaultInstance = new ServerLoad(true); + defaultInstance = new RegionInfo(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:ServerLoad) + // @@protoc_insertion_point(class_scope:RegionInfo) } - public interface TimeRangeOrBuilder + public interface FavoredNodesOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint64 from = 1; - boolean hasFrom(); - long getFrom(); - - // optional uint64 to = 2; - boolean hasTo(); - long getTo(); + // repeated .ServerName favoredNode = 1; + java.util.List + getFavoredNodeList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index); + int getFavoredNodeCount(); + java.util.List + getFavoredNodeOrBuilderList(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getFavoredNodeOrBuilder( + int index); } - public static final class TimeRange extends + public static final class FavoredNodes extends com.google.protobuf.GeneratedMessage - implements TimeRangeOrBuilder { - // Use TimeRange.newBuilder() to construct. - private TimeRange(Builder builder) { + implements FavoredNodesOrBuilder { + // Use FavoredNodes.newBuilder() to construct. + private FavoredNodes(Builder builder) { super(builder); } - private TimeRange(boolean noInit) {} + private FavoredNodes(boolean noInit) {} - private static final TimeRange defaultInstance; - public static TimeRange getDefaultInstance() { + private static final FavoredNodes defaultInstance; + public static FavoredNodes getDefaultInstance() { return defaultInstance; } - public TimeRange getDefaultInstanceForType() { + public FavoredNodes getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable; } - private int bitField0_; - // optional uint64 from = 1; - public static final int FROM_FIELD_NUMBER = 1; - private long from_; - public boolean hasFrom() { - return ((bitField0_ & 0x00000001) == 0x00000001); + // repeated .ServerName favoredNode = 1; + public static final int FAVOREDNODE_FIELD_NUMBER = 1; + private java.util.List favoredNode_; + public java.util.List getFavoredNodeList() { + return favoredNode_; } - public long getFrom() { - return from_; + public java.util.List + getFavoredNodeOrBuilderList() { + return favoredNode_; } - - // optional uint64 to = 2; - public static final int TO_FIELD_NUMBER = 2; - private long to_; - public boolean hasTo() { - return ((bitField0_ & 0x00000002) == 0x00000002); + public int getFavoredNodeCount() { + return favoredNode_.size(); } - public long getTo() { - return to_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index) { + return favoredNode_.get(index); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getFavoredNodeOrBuilder( + int index) { + return favoredNode_.get(index); } private void initFields() { - from_ = 0L; - to_ = 0L; + favoredNode_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; + for (int i = 0; i < getFavoredNodeCount(); i++) { + if (!getFavoredNode(i).isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeUInt64(1, from_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeUInt64(2, to_); + throws java.io.IOException { + getSerializedSize(); + for (int i = 0; i < favoredNode_.size(); i++) { + output.writeMessage(1, favoredNode_.get(i)); } getUnknownFields().writeTo(output); } @@ -7862,13 +3190,9 @@ public final class HBaseProtos { if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(1, from_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + for (int i = 0; i < favoredNode_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(2, to_); + .computeMessageSize(1, favoredNode_.get(i)); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -7887,22 +3211,14 @@ public final class HBaseProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange) obj; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes) obj; boolean result = true; - result = result && (hasFrom() == other.hasFrom()); - if (hasFrom()) { - result = result && (getFrom() - == other.getFrom()); - } - result = result && (hasTo() == other.hasTo()); - if (hasTo()) { - result = result && (getTo() - == other.getTo()); - } + result = result && getFavoredNodeList() + .equals(other.getFavoredNodeList()); result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -7912,53 +3228,49 @@ public final class HBaseProtos { public int hashCode() { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasFrom()) { - hash = (37 * hash) + FROM_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getFrom()); - } - if (hasTo()) { - hash = (37 * hash) + TO_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getTo()); + if (getFavoredNodeCount() > 0) { + hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER; + hash = (53 * hash) + getFavoredNodeList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -7967,7 +3279,7 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -7978,12 +3290,12 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -7993,7 +3305,7 @@ public final class HBaseProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -8006,18 +3318,18 @@ public final class HBaseProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodesOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -8028,6 +3340,7 @@ public final class HBaseProtos { } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + getFavoredNodeFieldBuilder(); } } private static Builder create() { @@ -8036,10 +3349,12 @@ public final class HBaseProtos { public Builder clear() { super.clear(); - from_ = 0L; - bitField0_ = (bitField0_ & ~0x00000001); - to_ = 0L; - bitField0_ = (bitField0_ & ~0x00000002); + if (favoredNodeBuilder_ == null) { + favoredNode_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000001); + } else { + favoredNodeBuilder_.clear(); + } return this; } @@ -8049,24 +3364,24 @@ public final class HBaseProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -8074,45 +3389,70 @@ public final class HBaseProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange(this); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes(this); int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.from_ = from_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; + if (favoredNodeBuilder_ == null) { + if (((bitField0_ & 0x00000001) == 0x00000001)) { + favoredNode_ = java.util.Collections.unmodifiableList(favoredNode_); + bitField0_ = (bitField0_ & ~0x00000001); + } + result.favoredNode_ = favoredNode_; + } else { + result.favoredNode_ = favoredNodeBuilder_.build(); } - result.to_ = to_; - result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance()) return this; - if (other.hasFrom()) { - setFrom(other.getFrom()); - } - if (other.hasTo()) { - setTo(other.getTo()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.getDefaultInstance()) return this; + if (favoredNodeBuilder_ == null) { + if (!other.favoredNode_.isEmpty()) { + if (favoredNode_.isEmpty()) { + favoredNode_ = other.favoredNode_; + bitField0_ = (bitField0_ & ~0x00000001); + } else { + ensureFavoredNodeIsMutable(); + favoredNode_.addAll(other.favoredNode_); + } + onChanged(); + } + } else { + if (!other.favoredNode_.isEmpty()) { + if (favoredNodeBuilder_.isEmpty()) { + favoredNodeBuilder_.dispose(); + favoredNodeBuilder_ = null; + favoredNode_ = other.favoredNode_; + bitField0_ = (bitField0_ & ~0x00000001); + favoredNodeBuilder_ = + com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + getFavoredNodeFieldBuilder() : null; + } else { + favoredNodeBuilder_.addAllMessages(other.favoredNode_); + } + } } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { + for (int i = 0; i < getFavoredNodeCount(); i++) { + if (!getFavoredNode(i).isInitialized()) { + + return false; + } + } return true; } @@ -8139,167 +3479,358 @@ public final class HBaseProtos { } break; } - case 8: { - bitField0_ |= 0x00000001; - from_ = input.readUInt64(); - break; - } - case 16: { - bitField0_ |= 0x00000002; - to_ = input.readUInt64(); + case 10: { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(); + input.readMessage(subBuilder, extensionRegistry); + addFavoredNode(subBuilder.buildPartial()); break; } } } } - - private int bitField0_; - - // optional uint64 from = 1; - private long from_ ; - public boolean hasFrom() { - return ((bitField0_ & 0x00000001) == 0x00000001); + + private int bitField0_; + + // repeated .ServerName favoredNode = 1; + private java.util.List favoredNode_ = + java.util.Collections.emptyList(); + private void ensureFavoredNodeIsMutable() { + if (!((bitField0_ & 0x00000001) == 0x00000001)) { + favoredNode_ = new java.util.ArrayList(favoredNode_); + bitField0_ |= 0x00000001; + } + } + + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> favoredNodeBuilder_; + + public java.util.List getFavoredNodeList() { + if (favoredNodeBuilder_ == null) { + return java.util.Collections.unmodifiableList(favoredNode_); + } else { + return favoredNodeBuilder_.getMessageList(); + } + } + public int getFavoredNodeCount() { + if (favoredNodeBuilder_ == null) { + return favoredNode_.size(); + } else { + return favoredNodeBuilder_.getCount(); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index) { + if (favoredNodeBuilder_ == null) { + return favoredNode_.get(index); + } else { + return favoredNodeBuilder_.getMessage(index); + } + } + public Builder setFavoredNode( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { + if (favoredNodeBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureFavoredNodeIsMutable(); + favoredNode_.set(index, value); + onChanged(); + } else { + favoredNodeBuilder_.setMessage(index, value); + } + return this; + } + public Builder setFavoredNode( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { + if (favoredNodeBuilder_ == null) { + ensureFavoredNodeIsMutable(); + favoredNode_.set(index, builderForValue.build()); + onChanged(); + } else { + favoredNodeBuilder_.setMessage(index, builderForValue.build()); + } + return this; + } + public Builder addFavoredNode(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { + if (favoredNodeBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureFavoredNodeIsMutable(); + favoredNode_.add(value); + onChanged(); + } else { + favoredNodeBuilder_.addMessage(value); + } + return this; + } + public Builder addFavoredNode( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { + if (favoredNodeBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + ensureFavoredNodeIsMutable(); + favoredNode_.add(index, value); + onChanged(); + } else { + favoredNodeBuilder_.addMessage(index, value); + } + return this; + } + public Builder addFavoredNode( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { + if (favoredNodeBuilder_ == null) { + ensureFavoredNodeIsMutable(); + favoredNode_.add(builderForValue.build()); + onChanged(); + } else { + favoredNodeBuilder_.addMessage(builderForValue.build()); + } + return this; } - public long getFrom() { - return from_; + public Builder addFavoredNode( + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { + if (favoredNodeBuilder_ == null) { + ensureFavoredNodeIsMutable(); + favoredNode_.add(index, builderForValue.build()); + onChanged(); + } else { + favoredNodeBuilder_.addMessage(index, builderForValue.build()); + } + return this; } - public Builder setFrom(long value) { - bitField0_ |= 0x00000001; - from_ = value; - onChanged(); + public Builder addAllFavoredNode( + java.lang.Iterable values) { + if (favoredNodeBuilder_ == null) { + ensureFavoredNodeIsMutable(); + super.addAll(values, favoredNode_); + onChanged(); + } else { + favoredNodeBuilder_.addAllMessages(values); + } return this; } - public Builder clearFrom() { - bitField0_ = (bitField0_ & ~0x00000001); - from_ = 0L; - onChanged(); + public Builder clearFavoredNode() { + if (favoredNodeBuilder_ == null) { + favoredNode_ = java.util.Collections.emptyList(); + bitField0_ = (bitField0_ & ~0x00000001); + onChanged(); + } else { + favoredNodeBuilder_.clear(); + } return this; } - - // optional uint64 to = 2; - private long to_ ; - public boolean hasTo() { - return ((bitField0_ & 0x00000002) == 0x00000002); + public Builder removeFavoredNode(int index) { + if (favoredNodeBuilder_ == null) { + ensureFavoredNodeIsMutable(); + favoredNode_.remove(index); + onChanged(); + } else { + favoredNodeBuilder_.remove(index); + } + return this; } - public long getTo() { - return to_; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getFavoredNodeBuilder( + int index) { + return getFavoredNodeFieldBuilder().getBuilder(index); } - public Builder setTo(long value) { - bitField0_ |= 0x00000002; - to_ = value; - onChanged(); - return this; + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getFavoredNodeOrBuilder( + int index) { + if (favoredNodeBuilder_ == null) { + return favoredNode_.get(index); } else { + return favoredNodeBuilder_.getMessageOrBuilder(index); + } } - public Builder clearTo() { - bitField0_ = (bitField0_ & ~0x00000002); - to_ = 0L; - onChanged(); - return this; + public java.util.List + getFavoredNodeOrBuilderList() { + if (favoredNodeBuilder_ != null) { + return favoredNodeBuilder_.getMessageOrBuilderList(); + } else { + return java.util.Collections.unmodifiableList(favoredNode_); + } + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder addFavoredNodeBuilder() { + return getFavoredNodeFieldBuilder().addBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder addFavoredNodeBuilder( + int index) { + return getFavoredNodeFieldBuilder().addBuilder( + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); + } + public java.util.List + getFavoredNodeBuilderList() { + return getFavoredNodeFieldBuilder().getBuilderList(); + } + private com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + getFavoredNodeFieldBuilder() { + if (favoredNodeBuilder_ == null) { + favoredNodeBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( + favoredNode_, + ((bitField0_ & 0x00000001) == 0x00000001), + getParentForChildren(), + isClean()); + favoredNode_ = null; + } + return favoredNodeBuilder_; } - // @@protoc_insertion_point(builder_scope:TimeRange) + // @@protoc_insertion_point(builder_scope:FavoredNodes) } static { - defaultInstance = new TimeRange(true); + defaultInstance = new FavoredNodes(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:TimeRange) + // @@protoc_insertion_point(class_scope:FavoredNodes) } - public interface FilterOrBuilder + public interface RegionSpecifierOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string name = 1; - boolean hasName(); - String getName(); + // required .RegionSpecifier.RegionSpecifierType type = 1; + boolean hasType(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType getType(); - // optional bytes serializedFilter = 2; - boolean hasSerializedFilter(); - com.google.protobuf.ByteString getSerializedFilter(); + // required bytes value = 2; + boolean hasValue(); + com.google.protobuf.ByteString getValue(); } - public static final class Filter extends + public static final class RegionSpecifier extends com.google.protobuf.GeneratedMessage - implements FilterOrBuilder { - // Use Filter.newBuilder() to construct. - private Filter(Builder builder) { + implements RegionSpecifierOrBuilder { + // Use RegionSpecifier.newBuilder() to construct. + private RegionSpecifier(Builder builder) { super(builder); } - private Filter(boolean noInit) {} + private RegionSpecifier(boolean noInit) {} - private static final Filter defaultInstance; - public static Filter getDefaultInstance() { + private static final RegionSpecifier defaultInstance; + public static RegionSpecifier getDefaultInstance() { return defaultInstance; } - public Filter getDefaultInstanceForType() { + public RegionSpecifier getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Filter_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Filter_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_fieldAccessorTable; } - private int bitField0_; - // required string name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private java.lang.Object name_; - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - public String getName() { - java.lang.Object ref = name_; - if (ref instanceof String) { - return (String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - String s = bs.toStringUtf8(); - if (com.google.protobuf.Internal.isValidUtf8(bs)) { - name_ = s; + public enum RegionSpecifierType + implements com.google.protobuf.ProtocolMessageEnum { + REGION_NAME(0, 1), + ENCODED_REGION_NAME(1, 2), + ; + + public static final int REGION_NAME_VALUE = 1; + public static final int ENCODED_REGION_NAME_VALUE = 2; + + + public final int getNumber() { return value; } + + public static RegionSpecifierType valueOf(int value) { + switch (value) { + case 1: return REGION_NAME; + case 2: return ENCODED_REGION_NAME; + default: return null; } - return s; } - } - private com.google.protobuf.ByteString getNameBytes() { - java.lang.Object ref = name_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8((String) ref); - name_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; + + public static com.google.protobuf.Internal.EnumLiteMap + internalGetValueMap() { + return internalValueMap; + } + private static com.google.protobuf.Internal.EnumLiteMap + internalValueMap = + new com.google.protobuf.Internal.EnumLiteMap() { + public RegionSpecifierType findValueByNumber(int number) { + return RegionSpecifierType.valueOf(number); + } + }; + + public final com.google.protobuf.Descriptors.EnumValueDescriptor + getValueDescriptor() { + return getDescriptor().getValues().get(index); + } + public final com.google.protobuf.Descriptors.EnumDescriptor + getDescriptorForType() { + return getDescriptor(); + } + public static final com.google.protobuf.Descriptors.EnumDescriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDescriptor().getEnumTypes().get(0); + } + + private static final RegionSpecifierType[] VALUES = { + REGION_NAME, ENCODED_REGION_NAME, + }; + + public static RegionSpecifierType valueOf( + com.google.protobuf.Descriptors.EnumValueDescriptor desc) { + if (desc.getType() != getDescriptor()) { + throw new java.lang.IllegalArgumentException( + "EnumValueDescriptor is not for this type."); + } + return VALUES[desc.getIndex()]; + } + + private final int index; + private final int value; + + private RegionSpecifierType(int index, int value) { + this.index = index; + this.value = value; } + + // @@protoc_insertion_point(enum_scope:RegionSpecifier.RegionSpecifierType) + } + + private int bitField0_; + // required .RegionSpecifier.RegionSpecifierType type = 1; + public static final int TYPE_FIELD_NUMBER = 1; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType type_; + public boolean hasType() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType getType() { + return type_; } - // optional bytes serializedFilter = 2; - public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString serializedFilter_; - public boolean hasSerializedFilter() { + // required bytes value = 2; + public static final int VALUE_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString value_; + public boolean hasValue() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public com.google.protobuf.ByteString getSerializedFilter() { - return serializedFilter_; + public com.google.protobuf.ByteString getValue() { + return value_; } private void initFields() { - name_ = ""; - serializedFilter_ = com.google.protobuf.ByteString.EMPTY; + type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; + value_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - if (!hasName()) { + if (!hasType()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasValue()) { memoizedIsInitialized = 0; return false; } @@ -8311,10 +3842,10 @@ public final class HBaseProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getNameBytes()); + output.writeEnum(1, type_.getNumber()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, serializedFilter_); + output.writeBytes(2, value_); } getUnknownFields().writeTo(output); } @@ -8327,11 +3858,11 @@ public final class HBaseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getNameBytes()); + .computeEnumSize(1, type_.getNumber()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, serializedFilter_); + .computeBytesSize(2, value_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -8350,21 +3881,21 @@ public final class HBaseProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter) obj; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier) obj; boolean result = true; - result = result && (hasName() == other.hasName()); - if (hasName()) { - result = result && getName() - .equals(other.getName()); + result = result && (hasType() == other.hasType()); + if (hasType()) { + result = result && + (getType() == other.getType()); } - result = result && (hasSerializedFilter() == other.hasSerializedFilter()); - if (hasSerializedFilter()) { - result = result && getSerializedFilter() - .equals(other.getSerializedFilter()); + result = result && (hasValue() == other.hasValue()); + if (hasValue()) { + result = result && getValue() + .equals(other.getValue()); } result = result && getUnknownFields().equals(other.getUnknownFields()); @@ -8375,53 +3906,53 @@ public final class HBaseProtos { public int hashCode() { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasName()) { - hash = (37 * hash) + NAME_FIELD_NUMBER; - hash = (53 * hash) + getName().hashCode(); + if (hasType()) { + hash = (37 * hash) + TYPE_FIELD_NUMBER; + hash = (53 * hash) + hashEnum(getType()); } - if (hasSerializedFilter()) { - hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER; - hash = (53 * hash) + getSerializedFilter().hashCode(); + if (hasValue()) { + hash = (37 * hash) + VALUE_FIELD_NUMBER; + hash = (53 * hash) + getValue().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -8430,7 +3961,7 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -8441,12 +3972,12 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -8456,7 +3987,7 @@ public final class HBaseProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -8469,18 +4000,18 @@ public final class HBaseProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FilterOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Filter_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_Filter_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_RegionSpecifier_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -8499,9 +4030,9 @@ public final class HBaseProtos { public Builder clear() { super.clear(); - name_ = ""; + type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; bitField0_ = (bitField0_ & ~0x00000001); - serializedFilter_ = com.google.protobuf.ByteString.EMPTY; + value_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } @@ -8512,24 +4043,24 @@ public final class HBaseProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -8537,46 +4068,50 @@ public final class HBaseProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter(this); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.name_ = name_; + result.type_ = type_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } - result.serializedFilter_ = serializedFilter_; + result.value_ = value_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.getDefaultInstance()) return this; - if (other.hasName()) { - setName(other.getName()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) return this; + if (other.hasType()) { + setType(other.getType()); } - if (other.hasSerializedFilter()) { - setSerializedFilter(other.getSerializedFilter()); + if (other.hasValue()) { + setValue(other.getValue()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - if (!hasName()) { + if (!hasType()) { + + return false; + } + if (!hasValue()) { return false; } @@ -8606,14 +4141,20 @@ public final class HBaseProtos { } break; } - case 10: { - bitField0_ |= 0x00000001; - name_ = input.readBytes(); + case 8: { + int rawValue = input.readEnum(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType value = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.valueOf(rawValue); + if (value == null) { + unknownFields.mergeVarintField(1, rawValue); + } else { + bitField0_ |= 0x00000001; + type_ = value; + } break; } case 18: { bitField0_ |= 0x00000002; - serializedFilter_ = input.readBytes(); + value_ = input.readBytes(); break; } } @@ -8622,218 +4163,134 @@ public final class HBaseProtos { private int bitField0_; - // required string name = 1; - private java.lang.Object name_ = ""; - public boolean hasName() { + // required .RegionSpecifier.RegionSpecifierType type = 1; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; + public boolean hasType() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public String getName() { - java.lang.Object ref = name_; - if (!(ref instanceof String)) { - String s = ((com.google.protobuf.ByteString) ref).toStringUtf8(); - name_ = s; - return s; - } else { - return (String) ref; - } + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType getType() { + return type_; } - public Builder setName(String value) { + public Builder setType(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType value) { if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + type_ = value; onChanged(); return this; } - public Builder clearName() { + public Builder clearType() { bitField0_ = (bitField0_ & ~0x00000001); - name_ = getDefaultInstance().getName(); + type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME; onChanged(); return this; } - void setName(com.google.protobuf.ByteString value) { - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - } - // optional bytes serializedFilter = 2; - private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasSerializedFilter() { + // required bytes value = 2; + private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; + public boolean hasValue() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public com.google.protobuf.ByteString getSerializedFilter() { - return serializedFilter_; + public com.google.protobuf.ByteString getValue() { + return value_; } - public Builder setSerializedFilter(com.google.protobuf.ByteString value) { + public Builder setValue(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; - serializedFilter_ = value; + value_ = value; onChanged(); return this; } - public Builder clearSerializedFilter() { + public Builder clearValue() { bitField0_ = (bitField0_ & ~0x00000002); - serializedFilter_ = getDefaultInstance().getSerializedFilter(); + value_ = getDefaultInstance().getValue(); onChanged(); return this; } - // @@protoc_insertion_point(builder_scope:Filter) + // @@protoc_insertion_point(builder_scope:RegionSpecifier) } static { - defaultInstance = new Filter(true); + defaultInstance = new RegionSpecifier(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:Filter) + // @@protoc_insertion_point(class_scope:RegionSpecifier) } - public interface KeyValueOrBuilder + public interface TimeRangeOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes row = 1; - boolean hasRow(); - com.google.protobuf.ByteString getRow(); - - // required bytes family = 2; - boolean hasFamily(); - com.google.protobuf.ByteString getFamily(); - - // required bytes qualifier = 3; - boolean hasQualifier(); - com.google.protobuf.ByteString getQualifier(); - - // optional uint64 timestamp = 4; - boolean hasTimestamp(); - long getTimestamp(); - - // optional .CellType keyType = 5; - boolean hasKeyType(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType getKeyType(); + // optional uint64 from = 1; + boolean hasFrom(); + long getFrom(); - // optional bytes value = 6; - boolean hasValue(); - com.google.protobuf.ByteString getValue(); + // optional uint64 to = 2; + boolean hasTo(); + long getTo(); } - public static final class KeyValue extends + public static final class TimeRange extends com.google.protobuf.GeneratedMessage - implements KeyValueOrBuilder { - // Use KeyValue.newBuilder() to construct. - private KeyValue(Builder builder) { + implements TimeRangeOrBuilder { + // Use TimeRange.newBuilder() to construct. + private TimeRange(Builder builder) { super(builder); } - private KeyValue(boolean noInit) {} + private TimeRange(boolean noInit) {} - private static final KeyValue defaultInstance; - public static KeyValue getDefaultInstance() { + private static final TimeRange defaultInstance; + public static TimeRange getDefaultInstance() { return defaultInstance; } - public KeyValue getDefaultInstanceForType() { + public TimeRange getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_KeyValue_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_KeyValue_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_fieldAccessorTable; } private int bitField0_; - // required bytes row = 1; - public static final int ROW_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString row_; - public boolean hasRow() { + // optional uint64 from = 1; + public static final int FROM_FIELD_NUMBER = 1; + private long from_; + public boolean hasFrom() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public com.google.protobuf.ByteString getRow() { - return row_; + public long getFrom() { + return from_; } - // required bytes family = 2; - public static final int FAMILY_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString family_; - public boolean hasFamily() { + // optional uint64 to = 2; + public static final int TO_FIELD_NUMBER = 2; + private long to_; + public boolean hasTo() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public com.google.protobuf.ByteString getFamily() { - return family_; - } - - // required bytes qualifier = 3; - public static final int QUALIFIER_FIELD_NUMBER = 3; - private com.google.protobuf.ByteString qualifier_; - public boolean hasQualifier() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public com.google.protobuf.ByteString getQualifier() { - return qualifier_; - } - - // optional uint64 timestamp = 4; - public static final int TIMESTAMP_FIELD_NUMBER = 4; - private long timestamp_; - public boolean hasTimestamp() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public long getTimestamp() { - return timestamp_; - } - - // optional .CellType keyType = 5; - public static final int KEYTYPE_FIELD_NUMBER = 5; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType keyType_; - public boolean hasKeyType() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType getKeyType() { - return keyType_; - } - - // optional bytes value = 6; - public static final int VALUE_FIELD_NUMBER = 6; - private com.google.protobuf.ByteString value_; - public boolean hasValue() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - public com.google.protobuf.ByteString getValue() { - return value_; + public long getTo() { + return to_; } private void initFields() { - row_ = com.google.protobuf.ByteString.EMPTY; - family_ = com.google.protobuf.ByteString.EMPTY; - qualifier_ = com.google.protobuf.ByteString.EMPTY; - timestamp_ = 0L; - keyType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - value_ = com.google.protobuf.ByteString.EMPTY; + from_ = 0L; + to_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - if (!hasRow()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasFamily()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasQualifier()) { - memoizedIsInitialized = 0; - return false; - } memoizedIsInitialized = 1; return true; } @@ -8842,22 +4299,10 @@ public final class HBaseProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, row_); + output.writeUInt64(1, from_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, family_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, qualifier_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeUInt64(4, timestamp_); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeEnum(5, keyType_.getNumber()); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, value_); + output.writeUInt64(2, to_); } getUnknownFields().writeTo(output); } @@ -8870,27 +4315,11 @@ public final class HBaseProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, row_); + .computeUInt64Size(1, from_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, family_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, qualifier_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeUInt64Size(4, timestamp_); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeEnumSize(5, keyType_.getNumber()); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, value_); + .computeUInt64Size(2, to_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -8909,41 +4338,21 @@ public final class HBaseProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue) obj; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange) obj; boolean result = true; - result = result && (hasRow() == other.hasRow()); - if (hasRow()) { - result = result && getRow() - .equals(other.getRow()); - } - result = result && (hasFamily() == other.hasFamily()); - if (hasFamily()) { - result = result && getFamily() - .equals(other.getFamily()); - } - result = result && (hasQualifier() == other.hasQualifier()); - if (hasQualifier()) { - result = result && getQualifier() - .equals(other.getQualifier()); - } - result = result && (hasTimestamp() == other.hasTimestamp()); - if (hasTimestamp()) { - result = result && (getTimestamp() - == other.getTimestamp()); - } - result = result && (hasKeyType() == other.hasKeyType()); - if (hasKeyType()) { - result = result && - (getKeyType() == other.getKeyType()); + result = result && (hasFrom() == other.hasFrom()); + if (hasFrom()) { + result = result && (getFrom() + == other.getFrom()); } - result = result && (hasValue() == other.hasValue()); - if (hasValue()) { - result = result && getValue() - .equals(other.getValue()); + result = result && (hasTo() == other.hasTo()); + if (hasTo()) { + result = result && (getTo() + == other.getTo()); } result = result && getUnknownFields().equals(other.getUnknownFields()); @@ -8954,69 +4363,53 @@ public final class HBaseProtos { public int hashCode() { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasRow()) { - hash = (37 * hash) + ROW_FIELD_NUMBER; - hash = (53 * hash) + getRow().hashCode(); - } - if (hasFamily()) { - hash = (37 * hash) + FAMILY_FIELD_NUMBER; - hash = (53 * hash) + getFamily().hashCode(); - } - if (hasQualifier()) { - hash = (37 * hash) + QUALIFIER_FIELD_NUMBER; - hash = (53 * hash) + getQualifier().hashCode(); - } - if (hasTimestamp()) { - hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getTimestamp()); - } - if (hasKeyType()) { - hash = (37 * hash) + KEYTYPE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getKeyType()); + if (hasFrom()) { + hash = (37 * hash) + FROM_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getFrom()); } - if (hasValue()) { - hash = (37 * hash) + VALUE_FIELD_NUMBER; - hash = (53 * hash) + getValue().hashCode(); + if (hasTo()) { + hash = (37 * hash) + TO_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getTo()); } hash = (29 * hash) + getUnknownFields().hashCode(); return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -9025,7 +4418,7 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -9036,12 +4429,12 @@ public final class HBaseProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -9051,7 +4444,7 @@ public final class HBaseProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -9064,18 +4457,18 @@ public final class HBaseProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValueOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_KeyValue_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_KeyValue_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TimeRange_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -9094,18 +4487,10 @@ public final class HBaseProtos { public Builder clear() { super.clear(); - row_ = com.google.protobuf.ByteString.EMPTY; + from_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); - family_ = com.google.protobuf.ByteString.EMPTY; + to_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); - qualifier_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000004); - timestamp_ = 0L; - bitField0_ = (bitField0_ & ~0x00000008); - keyType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - bitField0_ = (bitField0_ & ~0x00000010); - value_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000020); return this; } @@ -9115,24 +4500,24 @@ public final class HBaseProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -9140,85 +4525,45 @@ public final class HBaseProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue(this); + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.row_ = row_; + result.from_ = from_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } - result.family_ = family_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.qualifier_ = qualifier_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; - } - result.timestamp_ = timestamp_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; - } - result.keyType_ = keyType_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { - to_bitField0_ |= 0x00000020; - } - result.value_ = value_; + result.to_ = to_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue.getDefaultInstance()) return this; - if (other.hasRow()) { - setRow(other.getRow()); - } - if (other.hasFamily()) { - setFamily(other.getFamily()); - } - if (other.hasQualifier()) { - setQualifier(other.getQualifier()); - } - if (other.hasTimestamp()) { - setTimestamp(other.getTimestamp()); - } - if (other.hasKeyType()) { - setKeyType(other.getKeyType()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance()) return this; + if (other.hasFrom()) { + setFrom(other.getFrom()); } - if (other.hasValue()) { - setValue(other.getValue()); + if (other.hasTo()) { + setTo(other.getTo()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - if (!hasRow()) { - - return false; - } - if (!hasFamily()) { - - return false; - } - if (!hasQualifier()) { - - return false; - } return true; } @@ -9245,40 +4590,14 @@ public final class HBaseProtos { } break; } - case 10: { + case 8: { bitField0_ |= 0x00000001; - row_ = input.readBytes(); + from_ = input.readUInt64(); break; } - case 18: { + case 16: { bitField0_ |= 0x00000002; - family_ = input.readBytes(); - break; - } - case 26: { - bitField0_ |= 0x00000004; - qualifier_ = input.readBytes(); - break; - } - case 32: { - bitField0_ |= 0x00000008; - timestamp_ = input.readUInt64(); - break; - } - case 40: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType value = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(5, rawValue); - } else { - bitField0_ |= 0x00000010; - keyType_ = value; - } - break; - } - case 50: { - bitField0_ |= 0x00000020; - value_ = input.readBytes(); + to_ = input.readUInt64(); break; } } @@ -9287,156 +4606,57 @@ public final class HBaseProtos { private int bitField0_; - // required bytes row = 1; - private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasRow() { + // optional uint64 from = 1; + private long from_ ; + public boolean hasFrom() { return ((bitField0_ & 0x00000001) == 0x00000001); } - public com.google.protobuf.ByteString getRow() { - return row_; + public long getFrom() { + return from_; } - public Builder setRow(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - row_ = value; + public Builder setFrom(long value) { + bitField0_ |= 0x00000001; + from_ = value; onChanged(); return this; } - public Builder clearRow() { + public Builder clearFrom() { bitField0_ = (bitField0_ & ~0x00000001); - row_ = getDefaultInstance().getRow(); + from_ = 0L; onChanged(); return this; } - // required bytes family = 2; - private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasFamily() { + // optional uint64 to = 2; + private long to_ ; + public boolean hasTo() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public com.google.protobuf.ByteString getFamily() { - return family_; + public long getTo() { + return to_; } - public Builder setFamily(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - family_ = value; + public Builder setTo(long value) { + bitField0_ |= 0x00000002; + to_ = value; onChanged(); return this; } - public Builder clearFamily() { + public Builder clearTo() { bitField0_ = (bitField0_ & ~0x00000002); - family_ = getDefaultInstance().getFamily(); - onChanged(); - return this; - } - - // required bytes qualifier = 3; - private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasQualifier() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - public com.google.protobuf.ByteString getQualifier() { - return qualifier_; - } - public Builder setQualifier(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000004; - qualifier_ = value; - onChanged(); - return this; - } - public Builder clearQualifier() { - bitField0_ = (bitField0_ & ~0x00000004); - qualifier_ = getDefaultInstance().getQualifier(); - onChanged(); - return this; - } - - // optional uint64 timestamp = 4; - private long timestamp_ ; - public boolean hasTimestamp() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - public long getTimestamp() { - return timestamp_; - } - public Builder setTimestamp(long value) { - bitField0_ |= 0x00000008; - timestamp_ = value; - onChanged(); - return this; - } - public Builder clearTimestamp() { - bitField0_ = (bitField0_ & ~0x00000008); - timestamp_ = 0L; - onChanged(); - return this; - } - - // optional .CellType keyType = 5; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - public boolean hasKeyType() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType getKeyType() { - return keyType_; - } - public Builder setKeyType(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000010; - keyType_ = value; - onChanged(); - return this; - } - public Builder clearKeyType() { - bitField0_ = (bitField0_ & ~0x00000010); - keyType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CellType.MINIMUM; - onChanged(); - return this; - } - - // optional bytes value = 6; - private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; - public boolean hasValue() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - public com.google.protobuf.ByteString getValue() { - return value_; - } - public Builder setValue(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000020; - value_ = value; - onChanged(); - return this; - } - public Builder clearValue() { - bitField0_ = (bitField0_ & ~0x00000020); - value_ = getDefaultInstance().getValue(); + to_ = 0L; onChanged(); return this; } - // @@protoc_insertion_point(builder_scope:KeyValue) + // @@protoc_insertion_point(builder_scope:TimeRange) } static { - defaultInstance = new KeyValue(true); + defaultInstance = new TimeRange(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:KeyValue) + // @@protoc_insertion_point(class_scope:TimeRange) } public interface ServerNameOrBuilder @@ -14658,11 +9878,6 @@ public final class HBaseProtos { } private static com.google.protobuf.Descriptors.Descriptor - internal_static_Cell_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_Cell_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor internal_static_TableSchema_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable @@ -14688,31 +9903,11 @@ public final class HBaseProtos { com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_RegionSpecifier_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor - internal_static_RegionLoad_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_RegionLoad_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor - internal_static_ServerLoad_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_ServerLoad_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor internal_static_TimeRange_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TimeRange_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor - internal_static_Filter_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_Filter_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor - internal_static_KeyValue_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_KeyValue_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor internal_static_ServerName_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable @@ -14776,85 +9971,50 @@ public final class HBaseProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\013hbase.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fa" + - "mily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimesta" + - "mp\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n" + - "\005value\030\006 \001(\014\"\225\001\n\013TableSchema\022\014\n\004name\030\001 \001" + - "(\014\022#\n\nattributes\030\002 \003(\0132\017.BytesBytesPair\022" + - "+\n\016columnFamilies\030\003 \003(\0132\023.ColumnFamilySc" + - "hema\022&\n\rconfiguration\030\004 \003(\0132\017.NameString" + - "Pair\"o\n\022ColumnFamilySchema\022\014\n\004name\030\001 \002(\014" + - "\022#\n\nattributes\030\002 \003(\0132\017.BytesBytesPair\022&\n" + - "\rconfiguration\030\003 \003(\0132\017.NameStringPair\"\207\001", - "\n\nRegionInfo\022\020\n\010regionId\030\001 \002(\004\022\021\n\ttableN" + - "ame\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014\022\016\n\006endKey\030\004 " + - "\001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005split\030\006 \001(\010\022\022\n\nr" + - "ecovering\030\007 \001(\010\"0\n\014FavoredNodes\022 \n\013favor" + - "edNode\030\001 \003(\0132\013.ServerName\"\225\001\n\017RegionSpec" + - "ifier\0222\n\004type\030\001 \002(\0162$.RegionSpecifier.Re" + - "gionSpecifierType\022\r\n\005value\030\002 \002(\014\"?\n\023Regi" + - "onSpecifierType\022\017\n\013REGION_NAME\020\001\022\027\n\023ENCO" + - "DED_REGION_NAME\020\002\"\260\003\n\nRegionLoad\022)\n\017regi" + - "onSpecifier\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006s", - "tores\030\002 \001(\r\022\022\n\nstorefiles\030\003 \001(\r\022\037\n\027store" + - "UncompressedSizeMB\030\004 \001(\r\022\027\n\017storefileSiz" + - "eMB\030\005 \001(\r\022\026\n\016memstoreSizeMB\030\006 \001(\r\022\034\n\024sto" + - "refileIndexSizeMB\030\007 \001(\r\022\031\n\021readRequestsC" + - "ount\030\010 \001(\004\022\032\n\022writeRequestsCount\030\t \001(\004\022\032" + - "\n\022totalCompactingKVs\030\n \001(\004\022\033\n\023currentCom" + - "pactedKVs\030\013 \001(\004\022\027\n\017rootIndexSizeKB\030\014 \001(\r" + - "\022\036\n\026totalStaticIndexSizeKB\030\r \001(\r\022\036\n\026tota" + - "lStaticBloomSizeKB\030\016 \001(\r\022\032\n\022completeSequ" + - "enceId\030\017 \001(\004\"\372\001\n\nServerLoad\022\030\n\020numberOfR", - "equests\030\001 \001(\r\022\035\n\025totalNumberOfRequests\030\002" + - " \001(\r\022\022\n\nusedHeapMB\030\003 \001(\r\022\021\n\tmaxHeapMB\030\004 " + - "\001(\r\022 \n\013regionLoads\030\005 \003(\0132\013.RegionLoad\022\"\n" + - "\014coprocessors\030\006 \003(\0132\014.Coprocessor\022\027\n\017rep" + - "ortStartTime\030\007 \001(\004\022\025\n\rreportEndTime\030\010 \001(" + - "\004\022\026\n\016infoServerPort\030\t \001(\r\"%\n\tTimeRange\022\014" + - "\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"0\n\006Filter\022\014\n\004na" + - "me\030\001 \002(\t\022\030\n\020serializedFilter\030\002 \001(\014\"x\n\010Ke" + - "yValue\022\013\n\003row\030\001 \002(\014\022\016\n\006family\030\002 \002(\014\022\021\n\tq" + - "ualifier\030\003 \002(\014\022\021\n\ttimestamp\030\004 \001(\004\022\032\n\007key", - "Type\030\005 \001(\0162\t.CellType\022\r\n\005value\030\006 \001(\014\"?\n\n" + - "ServerName\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001" + - "(\r\022\021\n\tstartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004" + - "name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002" + - "(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004na" + - "me\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPai" + - "r\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rName" + - "Int64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255" + - "\001\n\023SnapshotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005" + - "table\030\002 \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004", - "type\030\004 \001(\0162\031.SnapshotDescription.Type:\005F" + - "LUSH\022\017\n\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLE" + - "D\020\000\022\t\n\005FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n" + - "\007longMsg\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigde" + - "cimalMsg\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001" + - " \002(\004\022\023\n\013mostSigBits\030\002 \002(\004*`\n\010CellType\022\013\n" + - "\007MINIMUM\020\000\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELET" + - "E_COLUMN\020\014\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM" + - "\020\377\001*r\n\013CompareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_" + - "EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GRE", - "ATER_OR_EQUAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B" + - ">\n*org.apache.hadoop.hbase.protobuf.gene" + - "ratedB\013HBaseProtosH\001\240\001\001" + "\n\013hbase.proto\032\nCell.proto\"\225\001\n\013TableSchem" + + "a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" + + "tesBytesPair\022+\n\016columnFamilies\030\003 \003(\0132\023.C" + + "olumnFamilySchema\022&\n\rconfiguration\030\004 \003(\013" + + "2\017.NameStringPair\"o\n\022ColumnFamilySchema\022" + + "\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byte" + + "sBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Name" + + "StringPair\"\207\001\n\nRegionInfo\022\020\n\010regionId\030\001 " + + "\002(\004\022\021\n\ttableName\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014" + + "\022\016\n\006endKey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl", + "it\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"0\n\014FavoredN" + + "odes\022 \n\013favoredNode\030\001 \003(\0132\013.ServerName\"\225" + + "\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$.Regio" + + "nSpecifier.RegionSpecifierType\022\r\n\005value\030" + + "\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REGION_N" + + "AME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tTimeRa" + + "nge\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"?\n\nServerN" + + "ame\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\021\n\ts" + + "tartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 " + + "\002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005v", + "alue\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(" + + "\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005fi" + + "rst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pa" + + "ir\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255\001\n\023Snap" + + "shotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002" + + " \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004type\030\004 " + + "\001(\0162\031.SnapshotDescription.Type:\005FLUSH\022\017\n" + + "\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" + + "FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n\007longMs" + + "g\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigdecimalMs", + "g\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001 \002(\004\022\023\n" + + "\013mostSigBits\030\002 \002(\004*r\n\013CompareType\022\010\n\004LES" + + "S\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT" + + "_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREATE" + + "R\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop.hbas" + + "e.protobuf.generatedB\013HBaseProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; - internal_static_Cell_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_Cell_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_Cell_descriptor, - new java.lang.String[] { "Row", "Family", "Qualifier", "Timestamp", "CellType", "Value", }, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.class, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Cell.Builder.class); internal_static_TableSchema_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(0); internal_static_TableSchema_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TableSchema_descriptor, @@ -14862,7 +10022,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder.class); internal_static_ColumnFamilySchema_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(1); internal_static_ColumnFamilySchema_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ColumnFamilySchema_descriptor, @@ -14870,7 +10030,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder.class); internal_static_RegionInfo_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(2); internal_static_RegionInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RegionInfo_descriptor, @@ -14878,7 +10038,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder.class); internal_static_FavoredNodes_descriptor = - getDescriptor().getMessageTypes().get(4); + getDescriptor().getMessageTypes().get(3); internal_static_FavoredNodes_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FavoredNodes_descriptor, @@ -14886,55 +10046,23 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.FavoredNodes.Builder.class); internal_static_RegionSpecifier_descriptor = - getDescriptor().getMessageTypes().get(5); + getDescriptor().getMessageTypes().get(4); internal_static_RegionSpecifier_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RegionSpecifier_descriptor, new java.lang.String[] { "Type", "Value", }, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder.class); - internal_static_RegionLoad_descriptor = - getDescriptor().getMessageTypes().get(6); - internal_static_RegionLoad_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_RegionLoad_descriptor, - new java.lang.String[] { "RegionSpecifier", "Stores", "Storefiles", "StoreUncompressedSizeMB", "StorefileSizeMB", "MemstoreSizeMB", "StorefileIndexSizeMB", "ReadRequestsCount", "WriteRequestsCount", "TotalCompactingKVs", "CurrentCompactedKVs", "RootIndexSizeKB", "TotalStaticIndexSizeKB", "TotalStaticBloomSizeKB", "CompleteSequenceId", }, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.class, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad.Builder.class); - internal_static_ServerLoad_descriptor = - getDescriptor().getMessageTypes().get(7); - internal_static_ServerLoad_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_ServerLoad_descriptor, - new java.lang.String[] { "NumberOfRequests", "TotalNumberOfRequests", "UsedHeapMB", "MaxHeapMB", "RegionLoads", "Coprocessors", "ReportStartTime", "ReportEndTime", "InfoServerPort", }, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.class, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder.class); internal_static_TimeRange_descriptor = - getDescriptor().getMessageTypes().get(8); + getDescriptor().getMessageTypes().get(5); internal_static_TimeRange_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TimeRange_descriptor, new java.lang.String[] { "From", "To", }, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.Builder.class); - internal_static_Filter_descriptor = - getDescriptor().getMessageTypes().get(9); - internal_static_Filter_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_Filter_descriptor, - new java.lang.String[] { "Name", "SerializedFilter", }, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.class, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Filter.Builder.class); - internal_static_KeyValue_descriptor = - getDescriptor().getMessageTypes().get(10); - internal_static_KeyValue_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_KeyValue_descriptor, - new java.lang.String[] { "Row", "Family", "Qualifier", "Timestamp", "KeyType", "Value", }, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue.class, - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.KeyValue.Builder.class); internal_static_ServerName_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(6); internal_static_ServerName_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ServerName_descriptor, @@ -14942,7 +10070,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder.class); internal_static_Coprocessor_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(7); internal_static_Coprocessor_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_Coprocessor_descriptor, @@ -14950,7 +10078,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor.Builder.class); internal_static_NameStringPair_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(8); internal_static_NameStringPair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_NameStringPair_descriptor, @@ -14958,7 +10086,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder.class); internal_static_NameBytesPair_descriptor = - getDescriptor().getMessageTypes().get(14); + getDescriptor().getMessageTypes().get(9); internal_static_NameBytesPair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_NameBytesPair_descriptor, @@ -14966,7 +10094,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.Builder.class); internal_static_BytesBytesPair_descriptor = - getDescriptor().getMessageTypes().get(15); + getDescriptor().getMessageTypes().get(10); internal_static_BytesBytesPair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_BytesBytesPair_descriptor, @@ -14974,7 +10102,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder.class); internal_static_NameInt64Pair_descriptor = - getDescriptor().getMessageTypes().get(16); + getDescriptor().getMessageTypes().get(11); internal_static_NameInt64Pair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_NameInt64Pair_descriptor, @@ -14982,7 +10110,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameInt64Pair.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameInt64Pair.Builder.class); internal_static_SnapshotDescription_descriptor = - getDescriptor().getMessageTypes().get(17); + getDescriptor().getMessageTypes().get(12); internal_static_SnapshotDescription_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SnapshotDescription_descriptor, @@ -14990,7 +10118,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder.class); internal_static_EmptyMsg_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(13); internal_static_EmptyMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_EmptyMsg_descriptor, @@ -14998,7 +10126,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.Builder.class); internal_static_LongMsg_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(14); internal_static_LongMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_LongMsg_descriptor, @@ -15006,7 +10134,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.Builder.class); internal_static_BigDecimalMsg_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(15); internal_static_BigDecimalMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_BigDecimalMsg_descriptor, @@ -15014,7 +10142,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BigDecimalMsg.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BigDecimalMsg.Builder.class); internal_static_UUID_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(16); internal_static_UUID_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_UUID_descriptor, @@ -15027,6 +10155,7 @@ public final class HBaseProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.CellProtos.getDescriptor(), }, assigner); } diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java index 89eebc4..c9f01a4 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java @@ -1138,8 +1138,8 @@ public final class RegionServerStatusProtos { // optional .ServerLoad load = 2; boolean hasLoad(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getLoad(); - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder getLoadOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getLoad(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getLoadOrBuilder(); } public static final class RegionServerReportRequest extends com.google.protobuf.GeneratedMessage @@ -1185,20 +1185,20 @@ public final class RegionServerStatusProtos { // optional .ServerLoad load = 2; public static final int LOAD_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad load_; + private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad load_; public boolean hasLoad() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getLoad() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getLoad() { return load_; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder getLoadOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getLoadOrBuilder() { return load_; } private void initFields() { server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); - load_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + load_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -1424,7 +1424,7 @@ public final class RegionServerStatusProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (loadBuilder_ == null) { - load_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + load_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); } else { loadBuilder_.clear(); } @@ -1560,7 +1560,7 @@ public final class RegionServerStatusProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.newBuilder(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.newBuilder(); if (hasLoad()) { subBuilder.mergeFrom(getLoad()); } @@ -1665,20 +1665,20 @@ public final class RegionServerStatusProtos { } // optional .ServerLoad load = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad load_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad load_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder> loadBuilder_; + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder> loadBuilder_; public boolean hasLoad() { return ((bitField0_ & 0x00000002) == 0x00000002); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad getLoad() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getLoad() { if (loadBuilder_ == null) { return load_; } else { return loadBuilder_.getMessage(); } } - public Builder setLoad(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad value) { + public Builder setLoad(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad value) { if (loadBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -1692,7 +1692,7 @@ public final class RegionServerStatusProtos { return this; } public Builder setLoad( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder builderForValue) { if (loadBuilder_ == null) { load_ = builderForValue.build(); onChanged(); @@ -1702,12 +1702,12 @@ public final class RegionServerStatusProtos { bitField0_ |= 0x00000002; return this; } - public Builder mergeLoad(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad value) { + public Builder mergeLoad(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad value) { if (loadBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - load_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance()) { + load_ != org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance()) { load_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.newBuilder(load_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.newBuilder(load_).mergeFrom(value).buildPartial(); } else { load_ = value; } @@ -1720,7 +1720,7 @@ public final class RegionServerStatusProtos { } public Builder clearLoad() { if (loadBuilder_ == null) { - load_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.getDefaultInstance(); + load_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); onChanged(); } else { loadBuilder_.clear(); @@ -1728,12 +1728,12 @@ public final class RegionServerStatusProtos { bitField0_ = (bitField0_ & ~0x00000002); return this; } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder getLoadBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder getLoadBuilder() { bitField0_ |= 0x00000002; onChanged(); return getLoadFieldBuilder().getBuilder(); } - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder getLoadOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getLoadOrBuilder() { if (loadBuilder_ != null) { return loadBuilder_.getMessageOrBuilder(); } else { @@ -1741,11 +1741,11 @@ public final class RegionServerStatusProtos { } } private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder> getLoadFieldBuilder() { if (loadBuilder_ == null) { loadBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerLoadOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder>( load_, getParentForChildren(), isClean()); @@ -4175,31 +4175,32 @@ public final class RegionServerStatusProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\030RegionServerStatus.proto\032\013hbase.proto\"" + - "^\n\032RegionServerStartupRequest\022\014\n\004port\030\001 " + - "\002(\r\022\027\n\017serverStartCode\030\002 \002(\004\022\031\n\021serverCu" + - "rrentTime\030\003 \002(\004\"B\n\033RegionServerStartupRe" + - "sponse\022#\n\nmapEntries\030\001 \003(\0132\017.NameStringP" + - "air\"S\n\031RegionServerReportRequest\022\033\n\006serv" + - "er\030\001 \002(\0132\013.ServerName\022\031\n\004load\030\002 \001(\0132\013.Se" + - "rverLoad\"\034\n\032RegionServerReportResponse\"N" + - "\n\031ReportRSFatalErrorRequest\022\033\n\006server\030\001 " + - "\002(\0132\013.ServerName\022\024\n\014errorMessage\030\002 \002(\t\"\034", - "\n\032ReportRSFatalErrorResponse\"5\n\037GetLastF" + - "lushedSequenceIdRequest\022\022\n\nregionName\030\001 " + - "\002(\014\"A\n GetLastFlushedSequenceIdResponse\022" + - "\035\n\025lastFlushedSequenceId\030\001 \002(\0042\354\002\n\031Regio" + - "nServerStatusService\022P\n\023regionServerStar" + - "tup\022\033.RegionServerStartupRequest\032\034.Regio" + - "nServerStartupResponse\022M\n\022regionServerRe" + - "port\022\032.RegionServerReportRequest\032\033.Regio" + - "nServerReportResponse\022M\n\022reportRSFatalEr" + - "ror\022\032.ReportRSFatalErrorRequest\032\033.Report", - "RSFatalErrorResponse\022_\n\030getLastFlushedSe" + - "quenceId\022 .GetLastFlushedSequenceIdReque" + - "st\032!.GetLastFlushedSequenceIdResponseBN\n" + - "*org.apache.hadoop.hbase.protobuf.genera" + - "tedB\030RegionServerStatusProtosH\001\210\001\001\240\001\001" + "\n\030RegionServerStatus.proto\032\013hbase.proto\032" + + "\023ClusterStatus.proto\"^\n\032RegionServerStar" + + "tupRequest\022\014\n\004port\030\001 \002(\r\022\027\n\017serverStartC" + + "ode\030\002 \002(\004\022\031\n\021serverCurrentTime\030\003 \002(\004\"B\n\033" + + "RegionServerStartupResponse\022#\n\nmapEntrie" + + "s\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionServer" + + "ReportRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa" + + "me\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032RegionS" + + "erverReportResponse\"N\n\031ReportRSFatalErro" + + "rRequest\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\024\n", + "\014errorMessage\030\002 \002(\t\"\034\n\032ReportRSFatalErro" + + "rResponse\"5\n\037GetLastFlushedSequenceIdReq" + + "uest\022\022\n\nregionName\030\001 \002(\014\"A\n GetLastFlush" + + "edSequenceIdResponse\022\035\n\025lastFlushedSeque" + + "nceId\030\001 \002(\0042\354\002\n\031RegionServerStatusServic" + + "e\022P\n\023regionServerStartup\022\033.RegionServerS" + + "tartupRequest\032\034.RegionServerStartupRespo" + + "nse\022M\n\022regionServerReport\022\032.RegionServer" + + "ReportRequest\032\033.RegionServerReportRespon" + + "se\022M\n\022reportRSFatalError\022\032.ReportRSFatal", + "ErrorRequest\032\033.ReportRSFatalErrorRespons" + + "e\022_\n\030getLastFlushedSequenceId\022 .GetLastF" + + "lushedSequenceIdRequest\032!.GetLastFlushed" + + "SequenceIdResponseBN\n*org.apache.hadoop." + + "hbase.protobuf.generatedB\030RegionServerSt" + + "atusProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -4277,6 +4278,7 @@ public final class RegionServerStatusProtos { .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(), }, assigner); } diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java index 2c14bd2..3a1a212 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java @@ -8,7 +8,7 @@ public final class ZooKeeperProtos { public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } - public interface RootRegionServerOrBuilder + public interface MetaRegionServerOrBuilder extends com.google.protobuf.MessageOrBuilder { // required .ServerName server = 1; @@ -16,32 +16,32 @@ public final class ZooKeeperProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); } - public static final class RootRegionServer extends + public static final class MetaRegionServer extends com.google.protobuf.GeneratedMessage - implements RootRegionServerOrBuilder { - // Use RootRegionServer.newBuilder() to construct. - private RootRegionServer(Builder builder) { + implements MetaRegionServerOrBuilder { + // Use MetaRegionServer.newBuilder() to construct. + private MetaRegionServer(Builder builder) { super(builder); } - private RootRegionServer(boolean noInit) {} + private MetaRegionServer(boolean noInit) {} - private static final RootRegionServer defaultInstance; - public static RootRegionServer getDefaultInstance() { + private static final MetaRegionServer defaultInstance; + public static MetaRegionServer getDefaultInstance() { return defaultInstance; } - public RootRegionServer getDefaultInstanceForType() { + public MetaRegionServer getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RootRegionServer_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RootRegionServer_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_fieldAccessorTable; } private int bitField0_; @@ -114,10 +114,10 @@ public final class ZooKeeperProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer) obj; + org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer) obj; boolean result = true; result = result && (hasServer() == other.hasServer()); @@ -142,41 +142,41 @@ public final class ZooKeeperProtos { return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { @@ -185,7 +185,7 @@ public final class ZooKeeperProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -196,12 +196,12 @@ public final class ZooKeeperProtos { return null; } } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } - public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -211,7 +211,7 @@ public final class ZooKeeperProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -224,18 +224,18 @@ public final class ZooKeeperProtos { } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServerOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServerOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RootRegionServer_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RootRegionServer_fieldAccessorTable; + return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_fieldAccessorTable; } - // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -270,24 +270,24 @@ public final class ZooKeeperProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer.getDescriptor(); + return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.getDescriptor(); } - public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer build() { - org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer build() { + org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer buildParsed() + private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { - org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer result = buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); @@ -295,8 +295,8 @@ public final class ZooKeeperProtos { return result; } - public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer(this); + public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { @@ -313,16 +313,16 @@ public final class ZooKeeperProtos { } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer.getDefaultInstance()) return this; + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.getDefaultInstance()) return this; if (other.hasServer()) { mergeServer(other.getServer()); } @@ -470,15 +470,15 @@ public final class ZooKeeperProtos { return serverBuilder_; } - // @@protoc_insertion_point(builder_scope:RootRegionServer) + // @@protoc_insertion_point(builder_scope:MetaRegionServer) } static { - defaultInstance = new RootRegionServer(true); + defaultInstance = new MetaRegionServer(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:RootRegionServer) + // @@protoc_insertion_point(class_scope:MetaRegionServer) } public interface MasterOrBuilder @@ -6868,10 +6868,10 @@ public final class ZooKeeperProtos { } private static com.google.protobuf.Descriptors.Descriptor - internal_static_RootRegionServer_descriptor; + internal_static_MetaRegionServer_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_RootRegionServer_fieldAccessorTable; + internal_static_MetaRegionServer_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_Master_descriptor; private static @@ -6941,7 +6941,7 @@ public final class ZooKeeperProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\017ZooKeeper.proto\032\013hbase.proto\"/\n\020RootRe" + + "\n\017ZooKeeper.proto\032\013hbase.proto\"/\n\020MetaRe" + "gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\"" + "%\n\006Master\022\033\n\006master\030\001 \002(\0132\013.ServerName\"\036" + "\n\tClusterUp\022\021\n\tstartDate\030\001 \002(\t\"\203\001\n\020Regio" + @@ -6976,14 +6976,14 @@ public final class ZooKeeperProtos { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; - internal_static_RootRegionServer_descriptor = + internal_static_MetaRegionServer_descriptor = getDescriptor().getMessageTypes().get(0); - internal_static_RootRegionServer_fieldAccessorTable = new + internal_static_MetaRegionServer_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_RootRegionServer_descriptor, + internal_static_MetaRegionServer_descriptor, new java.lang.String[] { "Server", }, - org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer.class, - org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RootRegionServer.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.class, + org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.Builder.class); internal_static_Master_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_Master_fieldAccessorTable = new diff --git a/hbase-protocol/src/main/protobuf/Cell.proto b/hbase-protocol/src/main/protobuf/Cell.proto new file mode 100644 index 0000000..85f6eeb --- /dev/null +++ b/hbase-protocol/src/main/protobuf/Cell.proto @@ -0,0 +1,64 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +// Cell and KeyValue protos + +option java_package = "org.apache.hadoop.hbase.protobuf.generated"; +option java_outer_classname = "CellProtos"; +option java_generate_equals_and_hash = true; +option optimize_for = SPEED; + +/** + * The type of the key in a Cell + */ +enum CellType { + MINIMUM = 0; + PUT = 4; + + DELETE = 8; + DELETE_COLUMN = 12; + DELETE_FAMILY = 14; + + // MAXIMUM is used when searching; you look from maximum on down. + MAXIMUM = 255; +} + +/** + * Protocol buffer version of Cell. + */ +message Cell { + optional bytes row = 1; + optional bytes family = 2; + optional bytes qualifier = 3; + optional uint64 timestamp = 4; + optional CellType cellType = 5; + optional bytes value = 6; +} + +/** + * Protocol buffer version of KeyValue. + * It doesn't have those transient parameters + */ +message KeyValue { + required bytes row = 1; + required bytes family = 2; + required bytes qualifier = 3; + optional uint64 timestamp = 4; + optional CellType keyType = 5; + optional bytes value = 6; +} diff --git a/hbase-protocol/src/main/protobuf/Client.proto b/hbase-protocol/src/main/protobuf/Client.proto index 283c280..dcc9688 100644 --- a/hbase-protocol/src/main/protobuf/Client.proto +++ b/hbase-protocol/src/main/protobuf/Client.proto @@ -25,6 +25,8 @@ option java_generate_equals_and_hash = true; option optimize_for = SPEED; import "hbase.proto"; +import "Filter.proto"; +import "Cell.proto"; import "Comparator.proto"; /** diff --git a/hbase-protocol/src/main/protobuf/ClusterStatus.proto b/hbase-protocol/src/main/protobuf/ClusterStatus.proto index 49f8ffe..a8e2ac1 100644 --- a/hbase-protocol/src/main/protobuf/ClusterStatus.proto +++ b/hbase-protocol/src/main/protobuf/ClusterStatus.proto @@ -51,11 +51,110 @@ message RegionInTransition { required RegionState regionState = 2; } +message RegionLoad { + /** the region specifier */ + required RegionSpecifier regionSpecifier = 1; + + /** the number of stores for the region */ + optional uint32 stores = 2; + + /** the number of storefiles for the region */ + optional uint32 storefiles = 3; + + /** the total size of the store files for the region, uncompressed, in MB */ + optional uint32 storeUncompressedSizeMB = 4; + + /** the current total size of the store files for the region, in MB */ + optional uint32 storefileSizeMB = 5; + + /** the current size of the memstore for the region, in MB */ + optional uint32 memstoreSizeMB = 6; + + /** + * The current total size of root-level store file indexes for the region, + * in MB. The same as {@link #rootIndexSizeKB} but in MB. + */ + optional uint32 storefileIndexSizeMB = 7; + + /** the current total read requests made to region */ + optional uint64 readRequestsCount = 8; + + /** the current total write requests made to region */ + optional uint64 writeRequestsCount = 9; + + /** the total compacting key values in currently running compaction */ + optional uint64 totalCompactingKVs = 10; + + /** the completed count of key values in currently running compaction */ + optional uint64 currentCompactedKVs = 11; + + /** The current total size of root-level indexes for the region, in KB. */ + optional uint32 rootIndexSizeKB = 12; + + /** The total size of all index blocks, not just the root level, in KB. */ + optional uint32 totalStaticIndexSizeKB = 13; + + /** + * The total size of all Bloom filter blocks, not just loaded into the + * block cache, in KB. + */ + optional uint32 totalStaticBloomSizeKB = 14; + + /** the most recent sequence Id from cache flush */ + optional uint64 completeSequenceId = 15; +} + +/* Server-level protobufs */ + +message ServerLoad { + /** Number of requests since last report. */ + optional uint32 numberOfRequests = 1; + + /** Total Number of requests from the start of the region server. */ + optional uint32 totalNumberOfRequests = 2; + + /** the amount of used heap, in MB. */ + optional uint32 usedHeapMB = 3; + + /** the maximum allowable size of the heap, in MB. */ + optional uint32 maxHeapMB = 4; + + /** Information on the load of individual regions. */ + repeated RegionLoad regionLoads = 5; + + /** + * Regionserver-level coprocessors, e.g., WALObserver implementations. + * Region-level coprocessors, on the other hand, are stored inside RegionLoad + * objects. + */ + repeated Coprocessor coprocessors = 6; + + /** + * Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests) + * time is measured as the difference, measured in milliseconds, between the current time + * and midnight, January 1, 1970 UTC. + */ + optional uint64 reportStartTime = 7; + + /** + * Time when report was generated. + * time is measured as the difference, measured in milliseconds, between the current time + * and midnight, January 1, 1970 UTC. + */ + optional uint64 reportEndTime = 8; + + /** + * The port number that this region server is hosing an info server on. + */ + optional uint32 infoServerPort = 9; +} + message LiveServerInfo { required ServerName server = 1; required ServerLoad serverLoad = 2; } + message ClusterStatus { optional HBaseVersionFileContent hbaseVersion = 1; repeated LiveServerInfo liveServers = 2; diff --git a/hbase-protocol/src/main/protobuf/Filter.proto b/hbase-protocol/src/main/protobuf/Filter.proto index ed59a00..618ef30 100644 --- a/hbase-protocol/src/main/protobuf/Filter.proto +++ b/hbase-protocol/src/main/protobuf/Filter.proto @@ -27,6 +27,11 @@ option optimize_for = SPEED; import "hbase.proto"; import "Comparator.proto"; +message Filter { + required string name = 1; + optional bytes serializedFilter = 2; +} + message ColumnCountGetFilter { required int32 limit = 1; } diff --git a/hbase-protocol/src/main/protobuf/RegionServerStatus.proto b/hbase-protocol/src/main/protobuf/RegionServerStatus.proto index 37c2c63..767749d 100644 --- a/hbase-protocol/src/main/protobuf/RegionServerStatus.proto +++ b/hbase-protocol/src/main/protobuf/RegionServerStatus.proto @@ -25,6 +25,7 @@ option java_generate_equals_and_hash = true; option optimize_for = SPEED; import "hbase.proto"; +import "ClusterStatus.proto"; message RegionServerStartupRequest { /** Port number this regionserver is up on */ diff --git a/hbase-protocol/src/main/protobuf/ZooKeeper.proto b/hbase-protocol/src/main/protobuf/ZooKeeper.proto index c137cb9..e14ef0e 100644 --- a/hbase-protocol/src/main/protobuf/ZooKeeper.proto +++ b/hbase-protocol/src/main/protobuf/ZooKeeper.proto @@ -28,10 +28,10 @@ option optimize_for = SPEED; import "hbase.proto"; /** - * Content of the root-region-server znode. + * Content of the meta-region-server znode. */ -message RootRegionServer { - // The ServerName hosting the root region currently. +message MetaRegionServer { + // The ServerName hosting the meta region currently. required ServerName server = 1; } diff --git a/hbase-protocol/src/main/protobuf/hbase.proto b/hbase-protocol/src/main/protobuf/hbase.proto index f91035d..8846b89 100644 --- a/hbase-protocol/src/main/protobuf/hbase.proto +++ b/hbase-protocol/src/main/protobuf/hbase.proto @@ -23,32 +23,7 @@ option java_outer_classname = "HBaseProtos"; option java_generate_equals_and_hash = true; option optimize_for = SPEED; -/** - * The type of the key in a Cell - */ -enum CellType { - MINIMUM = 0; - PUT = 4; - - DELETE = 8; - DELETE_COLUMN = 12; - DELETE_FAMILY = 14; - - // MAXIMUM is used when searching; you look from maximum on down. - MAXIMUM = 255; -} - -/** - * Protocol buffer version of Cell. - */ -message Cell { - optional bytes row = 1; - optional bytes family = 2; - optional bytes qualifier = 3; - optional uint64 timestamp = 4; - optional CellType cellType = 5; - optional bytes value = 6; -} +import "Cell.proto"; /** * Table Schema @@ -110,104 +85,6 @@ message RegionSpecifier { } } -message RegionLoad { - /** the region specifier */ - required RegionSpecifier regionSpecifier = 1; - - /** the number of stores for the region */ - optional uint32 stores = 2; - - /** the number of storefiles for the region */ - optional uint32 storefiles = 3; - - /** the total size of the store files for the region, uncompressed, in MB */ - optional uint32 storeUncompressedSizeMB = 4; - - /** the current total size of the store files for the region, in MB */ - optional uint32 storefileSizeMB = 5; - - /** the current size of the memstore for the region, in MB */ - optional uint32 memstoreSizeMB = 6; - - /** - * The current total size of root-level store file indexes for the region, - * in MB. The same as {@link #rootIndexSizeKB} but in MB. - */ - optional uint32 storefileIndexSizeMB = 7; - - /** the current total read requests made to region */ - optional uint64 readRequestsCount = 8; - - /** the current total write requests made to region */ - optional uint64 writeRequestsCount = 9; - - /** the total compacting key values in currently running compaction */ - optional uint64 totalCompactingKVs = 10; - - /** the completed count of key values in currently running compaction */ - optional uint64 currentCompactedKVs = 11; - - /** The current total size of root-level indexes for the region, in KB. */ - optional uint32 rootIndexSizeKB = 12; - - /** The total size of all index blocks, not just the root level, in KB. */ - optional uint32 totalStaticIndexSizeKB = 13; - - /** - * The total size of all Bloom filter blocks, not just loaded into the - * block cache, in KB. - */ - optional uint32 totalStaticBloomSizeKB = 14; - - /** the most recent sequence Id from cache flush */ - optional uint64 completeSequenceId = 15; -} - -/* Server-level protobufs */ - -message ServerLoad { - /** Number of requests since last report. */ - optional uint32 numberOfRequests = 1; - - /** Total Number of requests from the start of the region server. */ - optional uint32 totalNumberOfRequests = 2; - - /** the amount of used heap, in MB. */ - optional uint32 usedHeapMB = 3; - - /** the maximum allowable size of the heap, in MB. */ - optional uint32 maxHeapMB = 4; - - /** Information on the load of individual regions. */ - repeated RegionLoad regionLoads = 5; - - /** - * Regionserver-level coprocessors, e.g., WALObserver implementations. - * Region-level coprocessors, on the other hand, are stored inside RegionLoad - * objects. - */ - repeated Coprocessor coprocessors = 6; - - /** - * Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests) - * time is measured as the difference, measured in milliseconds, between the current time - * and midnight, January 1, 1970 UTC. - */ - optional uint64 reportStartTime = 7; - - /** - * Time when report was generated. - * time is measured as the difference, measured in milliseconds, between the current time - * and midnight, January 1, 1970 UTC. - */ - optional uint64 reportEndTime = 8; - - /** - * The port number that this region server is hosing an info server on. - */ - optional uint32 infoServerPort = 9; -} - /** * A range of time. Both from and to are Java time * stamp in milliseconds. If you don't specify a time @@ -219,11 +96,6 @@ message TimeRange { optional uint64 to = 2; } -message Filter { - required string name = 1; - optional bytes serializedFilter = 2; -} - /* Comparison operators */ enum CompareType { LESS = 0; @@ -236,19 +108,6 @@ enum CompareType { } /** - * Protocol buffer version of KeyValue. - * It doesn't have those transient parameters - */ -message KeyValue { - required bytes row = 1; - required bytes family = 2; - required bytes qualifier = 3; - optional uint64 timestamp = 4; - optional CellType keyType = 5; - optional bytes value = 6; -} - -/** * Protocol buffer version of ServerName */ message ServerName { diff --git a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.jamon b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.jamon index 1042734..7dbf8af 100644 --- a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.jamon +++ b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.jamon @@ -30,7 +30,7 @@ org.apache.hadoop.hbase.ServerName; org.apache.hadoop.hbase.HBaseConfiguration; org.apache.hadoop.hbase.protobuf.ProtobufUtil; org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo; -org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad; +org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad; <%if format.equals("json") %> <& ../common/TaskMonitorTmpl; filter = filter; format = "json" &> diff --git a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RegionListTmpl.jamon b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RegionListTmpl.jamon index eddb1c1..eff885b 100644 --- a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RegionListTmpl.jamon +++ b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/RegionListTmpl.jamon @@ -29,7 +29,7 @@ org.apache.hadoop.hbase.HBaseConfiguration; org.apache.hadoop.hbase.protobuf.ProtobufUtil; org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo; - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad; + org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad; <%if (onlineRegions != null && onlineRegions.size() > 0) %> diff --git a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/ServerMetricsTmpl.jamon b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/ServerMetricsTmpl.jamon index 22cf438..408d735 100644 --- a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/ServerMetricsTmpl.jamon +++ b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/regionserver/ServerMetricsTmpl.jamon @@ -29,7 +29,7 @@ org.apache.hadoop.hbase.ServerName; org.apache.hadoop.hbase.HBaseConfiguration; org.apache.hadoop.hbase.protobuf.ProtobufUtil; org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo; -org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad; +org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad; org.apache.hadoop.hbase.metrics.histogram.MetricsHistogram; org.apache.hadoop.util.StringUtils; com.yammer.metrics.stats.Snapshot; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java index d7183d9..250a461 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.codec.BaseDecoder; import org.apache.hadoop.hbase.codec.BaseEncoder; import org.apache.hadoop.hbase.codec.Codec; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.CellProtos; import com.google.protobuf.ByteString; import org.apache.hadoop.classification.InterfaceStability; @@ -47,7 +47,7 @@ public class MessageCodec implements Codec { @Override public void write(Cell cell) throws IOException { checkFlushed(); - HBaseProtos.Cell.Builder builder = HBaseProtos.Cell.newBuilder(); + CellProtos.Cell.Builder builder = CellProtos.Cell.newBuilder(); // This copies bytes from Cell to ByteString. I don't see anyway around the copy. // ByteString is final. builder.setRow(ByteString.copyFrom(cell.getRowArray(), cell.getRowOffset(), @@ -57,10 +57,10 @@ public class MessageCodec implements Codec { builder.setQualifier(ByteString.copyFrom(cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength())); builder.setTimestamp(cell.getTimestamp()); - builder.setCellType(HBaseProtos.CellType.valueOf(cell.getTypeByte())); + builder.setCellType(CellProtos.CellType.valueOf(cell.getTypeByte())); builder.setValue(ByteString.copyFrom(cell.getValueArray(), cell.getValueOffset(), cell.getValueLength())); - HBaseProtos.Cell pbcell = builder.build(); + CellProtos.Cell pbcell = builder.build(); pbcell.writeDelimitedTo(this.out); } } @@ -71,7 +71,7 @@ public class MessageCodec implements Codec { } protected Cell parseCell() throws IOException { - HBaseProtos.Cell pbcell = HBaseProtos.Cell.parseDelimitedFrom(this.in); + CellProtos.Cell pbcell = CellProtos.Cell.parseDelimitedFrom(this.in); return CellUtil.createCell(pbcell.getRow().toByteArray(), pbcell.getFamily().toByteArray(), pbcell.getQualifier().toByteArray(), pbcell.getTimestamp(), (byte)pbcell.getCellType().getNumber(), diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java index 0ed3635..61276c9 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java @@ -102,7 +102,7 @@ import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.RequestConverter; import org.apache.hadoop.hbase.protobuf.ResponseConverter; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; @@ -1244,7 +1244,7 @@ MasterServices, Server { public RegionServerReportResponse regionServerReport( RpcController controller, RegionServerReportRequest request) throws ServiceException { try { - HBaseProtos.ServerLoad sl = request.getLoad(); + ClusterStatusProtos.ServerLoad sl = request.getLoad(); this.serverManager.regionServerReport(ProtobufUtil.toServerName(request.getServer()), new ServerLoad(sl)); if (sl != null && this.metricsMaster != null) { // Up our metrics. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java index ee8f34b..ad83067 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java @@ -46,7 +46,6 @@ import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.client.RetriesExhaustedException; import org.apache.hadoop.hbase.exceptions.ClockOutOfSyncException; -import org.apache.hadoop.hbase.exceptions.PleaseHoldException; import org.apache.hadoop.hbase.exceptions.YouAreDeadException; import org.apache.hadoop.hbase.exceptions.ZooKeeperConnectionException; import org.apache.hadoop.hbase.master.handler.MetaServerShutdownHandler; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java index ed8f709..334754a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java @@ -170,10 +170,10 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Mut import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanRequest; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanResponse; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLoad; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.GetLastFlushedSequenceIdRequest; @@ -978,7 +978,7 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa void tryRegionServerReport(long reportStartTime, long reportEndTime) throws IOException { - HBaseProtos.ServerLoad sl = buildServerLoad(reportStartTime, reportEndTime); + ClusterStatusProtos.ServerLoad sl = buildServerLoad(reportStartTime, reportEndTime); try { RegionServerReportRequest.Builder request = RegionServerReportRequest.newBuilder(); ServerName sn = ServerName.parseVersionedServerName( @@ -1000,7 +1000,7 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa } } - HBaseProtos.ServerLoad buildServerLoad(long reportStartTime, long reportEndTime) { + ClusterStatusProtos.ServerLoad buildServerLoad(long reportStartTime, long reportEndTime) { // We're getting the MetricsRegionServerWrapper here because the wrapper computes requests // per second, and other metrics As long as metrics are part of ServerLoad it's best to use // the wrapper to compute those numbers in one place. @@ -1013,7 +1013,8 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa MemoryUsage memory = ManagementFactory.getMemoryMXBean().getHeapMemoryUsage(); - HBaseProtos.ServerLoad.Builder serverLoad = HBaseProtos.ServerLoad.newBuilder(); + ClusterStatusProtos.ServerLoad.Builder serverLoad = + ClusterStatusProtos.ServerLoad.newBuilder(); serverLoad.setNumberOfRequests((int) regionServerWrapper.getRequestsPerSecond()); serverLoad.setTotalNumberOfRequests((int) regionServerWrapper.getTotalRequestCount()); serverLoad.setUsedHeapMB((int)(memory.getUsed() / 1024 / 1024)); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java index d0e1b6c..fd08358 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerLoad.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase; import static org.junit.Assert.*; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.junit.Test; import org.junit.experimental.categories.Category; @@ -56,7 +57,7 @@ public class TestServerLoad { assertTrue(slToString.contains("coprocessors=[]")); } - private HBaseProtos.ServerLoad createServerLoadProto() { + private ClusterStatusProtos.ServerLoad createServerLoadProto() { HBaseProtos.RegionSpecifier rSpecOne = HBaseProtos.RegionSpecifier.newBuilder() .setType(HBaseProtos.RegionSpecifier.RegionSpecifierType.ENCODED_REGION_NAME) @@ -66,17 +67,18 @@ public class TestServerLoad { .setType(HBaseProtos.RegionSpecifier.RegionSpecifierType.ENCODED_REGION_NAME) .setValue(ByteString.copyFromUtf8("QWERTYUIOP")).build(); - HBaseProtos.RegionLoad rlOne = - HBaseProtos.RegionLoad.newBuilder().setRegionSpecifier(rSpecOne).setStores(10) + ClusterStatusProtos.RegionLoad rlOne = + ClusterStatusProtos.RegionLoad.newBuilder().setRegionSpecifier(rSpecOne).setStores(10) .setStorefiles(101).setStoreUncompressedSizeMB(106).setStorefileSizeMB(520) .setStorefileIndexSizeMB(42).setRootIndexSizeKB(201).build(); - HBaseProtos.RegionLoad rlTwo = - HBaseProtos.RegionLoad.newBuilder().setRegionSpecifier(rSpecTwo).setStores(3) + ClusterStatusProtos.RegionLoad rlTwo = + ClusterStatusProtos.RegionLoad.newBuilder().setRegionSpecifier(rSpecTwo).setStores(3) .setStorefiles(13).setStoreUncompressedSizeMB(23).setStorefileSizeMB(300) .setStorefileIndexSizeMB(40).setRootIndexSizeKB(303).build(); - HBaseProtos.ServerLoad sl = - HBaseProtos.ServerLoad.newBuilder().addRegionLoads(rlOne).addRegionLoads(rlTwo).build(); + ClusterStatusProtos.ServerLoad sl = + ClusterStatusProtos.ServerLoad.newBuilder().addRegionLoads(rlOne). + addRegionLoads(rlTwo).build(); return sl; } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestColumnPaginationFilter.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestColumnPaginationFilter.java index 8da9761..9e549c0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestColumnPaginationFilter.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestColumnPaginationFilter.java @@ -18,23 +18,17 @@ */ package org.apache.hadoop.hbase.filter; -import java.io.ByteArrayInputStream; -import java.io.ByteArrayOutputStream; -import java.io.DataInputStream; -import java.io.DataOutputStream; +import static org.junit.Assert.assertTrue; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.SmallTests; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.FilterProtos; import org.apache.hadoop.hbase.util.Bytes; - import org.junit.Before; import org.junit.Test; import org.junit.experimental.categories.Category; -import static org.junit.Assert.assertTrue; - /** * Test for the ColumnPaginationFilter, used mainly to test the successful serialization of the filter. * More test functionality can be found within {@link org.apache.hadoop.hbase.filter.TestFilter#testColumnPaginationFilter()} @@ -65,7 +59,7 @@ public class TestColumnPaginationFilter } private Filter serializationTest(Filter filter) throws Exception { - HBaseProtos.Filter filterProto = ProtobufUtil.toFilter(filter); + FilterProtos.Filter filterProto = ProtobufUtil.toFilter(filter); Filter newFilter = ProtobufUtil.toFilter(filterProto); return newFilter; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java index d10e24a..d67880b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.MediumTests; import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos; import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.test.MetricsAssertHelper; @@ -72,7 +72,7 @@ public class TestMasterMetrics { HRegionServer rs = cluster.getRegionServer(0); request.setServer(ProtobufUtil.toServerName(rs.getServerName())); - HBaseProtos.ServerLoad sl = HBaseProtos.ServerLoad.newBuilder() + ClusterStatusProtos.ServerLoad sl = ClusterStatusProtos.ServerLoad.newBuilder() .setTotalNumberOfRequests(10000) .build(); master.getMetrics().getMetricsSource().init();