diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java index 49cccfc..12b3d8c 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java @@ -6422,11 +6422,11 @@ public final class AccessControlProtos { "Permission\":\n\027CheckPermissionsRequest\022\037\n" + "\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" + "ermissionsResponse2\373\001\n\024AccessControlServ" + - "ice\022&\n\005grant\022\r.GrantRequest\032\016.GrantRespo", - "nse\022)\n\006revoke\022\016.RevokeRequest\032\017.RevokeRe" + - "sponse\022G\n\022getUserPermissions\022\027.UserPermi" + + "ice\022&\n\005Grant\022\r.GrantRequest\032\016.GrantRespo", + "nse\022)\n\006Revoke\022\016.RevokeRequest\032\017.RevokeRe" + + "sponse\022G\n\022GetUserPermissions\022\027.UserPermi" + "ssionsRequest\032\030.UserPermissionsResponse\022" + - "G\n\020checkPermissions\022\030.CheckPermissionsRe" + + "G\n\020CheckPermissions\022\030.CheckPermissionsRe" + "quest\032\031.CheckPermissionsResponseBI\n*org." + "apache.hadoop.hbase.protobuf.generatedB\023" + "AccessControlProtosH\001\210\001\001\240\001\001" diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AdminProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AdminProtos.java index a3bd954..2fad51a 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AdminProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AdminProtos.java @@ -16,7 +16,7 @@ public final class AdminProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); - // optional bool compactionState = 2; + // optional bool compaction_state = 2; boolean hasCompactionState(); boolean getCompactionState(); } @@ -62,8 +62,8 @@ public final class AdminProtos { return region_; } - // optional bool compactionState = 2; - public static final int COMPACTIONSTATE_FIELD_NUMBER = 2; + // optional bool compaction_state = 2; + public static final int COMPACTION_STATE_FIELD_NUMBER = 2; private boolean compactionState_; public boolean hasCompactionState() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -166,7 +166,7 @@ public final class AdminProtos { hash = (53 * hash) + getRegion().hashCode(); } if (hasCompactionState()) { - hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER; + hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getCompactionState()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -515,7 +515,7 @@ public final class AdminProtos { return regionBuilder_; } - // optional bool compactionState = 2; + // optional bool compaction_state = 2; private boolean compactionState_ ; public boolean hasCompactionState() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -550,12 +550,12 @@ public final class AdminProtos { public interface GetRegionInfoResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .RegionInfo regionInfo = 1; + // required .RegionInfo region_info = 1; boolean hasRegionInfo(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(); - // optional .GetRegionInfoResponse.CompactionState compactionState = 2; + // optional .GetRegionInfoResponse.CompactionState compaction_state = 2; boolean hasCompactionState(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState(); } @@ -663,8 +663,8 @@ public final class AdminProtos { } private int bitField0_; - // required .RegionInfo regionInfo = 1; - public static final int REGIONINFO_FIELD_NUMBER = 1; + // required .RegionInfo region_info = 1; + public static final int REGION_INFO_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_; public boolean hasRegionInfo() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -676,8 +676,8 @@ public final class AdminProtos { return regionInfo_; } - // optional .GetRegionInfoResponse.CompactionState compactionState = 2; - public static final int COMPACTIONSTATE_FIELD_NUMBER = 2; + // optional .GetRegionInfoResponse.CompactionState compaction_state = 2; + public static final int COMPACTION_STATE_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_; public boolean hasCompactionState() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -776,11 +776,11 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionInfo()) { - hash = (37 * hash) + REGIONINFO_FIELD_NUMBER; + hash = (37 * hash) + REGION_INFO_FIELD_NUMBER; hash = (53 * hash) + getRegionInfo().hashCode(); } if (hasCompactionState()) { - hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER; + hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getCompactionState()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1045,7 +1045,7 @@ public final class AdminProtos { private int bitField0_; - // required .RegionInfo regionInfo = 1; + // required .RegionInfo region_info = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_; @@ -1135,7 +1135,7 @@ public final class AdminProtos { return regionInfoBuilder_; } - // optional .GetRegionInfoResponse.CompactionState compactionState = 2; + // optional .GetRegionInfoResponse.CompactionState compaction_state = 2; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE; public boolean hasCompactionState() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1757,7 +1757,7 @@ public final class AdminProtos { public interface GetStoreFileResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated string storeFile = 1; + // repeated string store_file = 1; java.util.List getStoreFileList(); int getStoreFileCount(); String getStoreFile(int index); @@ -1790,8 +1790,8 @@ public final class AdminProtos { return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable; } - // repeated string storeFile = 1; - public static final int STOREFILE_FIELD_NUMBER = 1; + // repeated string store_file = 1; + public static final int STORE_FILE_FIELD_NUMBER = 1; private com.google.protobuf.LazyStringList storeFile_; public java.util.List getStoreFileList() { @@ -1875,7 +1875,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getStoreFileCount() > 0) { - hash = (37 * hash) + STOREFILE_FIELD_NUMBER; + hash = (37 * hash) + STORE_FILE_FIELD_NUMBER; hash = (53 * hash) + getStoreFileList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2106,7 +2106,7 @@ public final class AdminProtos { private int bitField0_; - // repeated string storeFile = 1; + // repeated string store_file = 1; private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureStoreFileIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { @@ -2476,7 +2476,7 @@ public final class AdminProtos { public interface GetOnlineRegionResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .RegionInfo regionInfo = 1; + // repeated .RegionInfo region_info = 1; java.util.List getRegionInfoList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index); @@ -2514,8 +2514,8 @@ public final class AdminProtos { return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable; } - // repeated .RegionInfo regionInfo = 1; - public static final int REGIONINFO_FIELD_NUMBER = 1; + // repeated .RegionInfo region_info = 1; + public static final int REGION_INFO_FIELD_NUMBER = 1; private java.util.List regionInfo_; public java.util.List getRegionInfoList() { return regionInfo_; @@ -2607,7 +2607,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getRegionInfoCount() > 0) { - hash = (37 * hash) + REGIONINFO_FIELD_NUMBER; + hash = (37 * hash) + REGION_INFO_FIELD_NUMBER; hash = (53 * hash) + getRegionInfoList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2869,7 +2869,7 @@ public final class AdminProtos { private int bitField0_; - // repeated .RegionInfo regionInfo = 1; + // repeated .RegionInfo region_info = 1; private java.util.List regionInfo_ = java.util.Collections.emptyList(); private void ensureRegionInfoIsMutable() { @@ -3069,7 +3069,7 @@ public final class AdminProtos { public interface OpenRegionRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1; + // repeated .OpenRegionRequest.RegionOpenInfo open_info = 1; java.util.List getOpenInfoList(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index); @@ -3115,11 +3115,11 @@ public final class AdminProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder(); - // optional uint32 versionOfOfflineNode = 2; + // optional uint32 version_of_offline_node = 2; boolean hasVersionOfOfflineNode(); int getVersionOfOfflineNode(); - // repeated .ServerName favoredNodes = 3; + // repeated .ServerName favored_nodes = 3; java.util.List getFavoredNodesList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index); @@ -3171,8 +3171,8 @@ public final class AdminProtos { return region_; } - // optional uint32 versionOfOfflineNode = 2; - public static final int VERSIONOFOFFLINENODE_FIELD_NUMBER = 2; + // optional uint32 version_of_offline_node = 2; + public static final int VERSION_OF_OFFLINE_NODE_FIELD_NUMBER = 2; private int versionOfOfflineNode_; public boolean hasVersionOfOfflineNode() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3181,8 +3181,8 @@ public final class AdminProtos { return versionOfOfflineNode_; } - // repeated .ServerName favoredNodes = 3; - public static final int FAVOREDNODES_FIELD_NUMBER = 3; + // repeated .ServerName favored_nodes = 3; + public static final int FAVORED_NODES_FIELD_NUMBER = 3; private java.util.List favoredNodes_; public java.util.List getFavoredNodesList() { return favoredNodes_; @@ -3312,11 +3312,11 @@ public final class AdminProtos { hash = (53 * hash) + getRegion().hashCode(); } if (hasVersionOfOfflineNode()) { - hash = (37 * hash) + VERSIONOFOFFLINENODE_FIELD_NUMBER; + hash = (37 * hash) + VERSION_OF_OFFLINE_NODE_FIELD_NUMBER; hash = (53 * hash) + getVersionOfOfflineNode(); } if (getFavoredNodesCount() > 0) { - hash = (37 * hash) + FAVOREDNODES_FIELD_NUMBER; + hash = (37 * hash) + FAVORED_NODES_FIELD_NUMBER; hash = (53 * hash) + getFavoredNodesList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3719,7 +3719,7 @@ public final class AdminProtos { return regionBuilder_; } - // optional uint32 versionOfOfflineNode = 2; + // optional uint32 version_of_offline_node = 2; private int versionOfOfflineNode_ ; public boolean hasVersionOfOfflineNode() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3740,7 +3740,7 @@ public final class AdminProtos { return this; } - // repeated .ServerName favoredNodes = 3; + // repeated .ServerName favored_nodes = 3; private java.util.List favoredNodes_ = java.util.Collections.emptyList(); private void ensureFavoredNodesIsMutable() { @@ -3937,8 +3937,8 @@ public final class AdminProtos { // @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo) } - // repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1; - public static final int OPENINFO_FIELD_NUMBER = 1; + // repeated .OpenRegionRequest.RegionOpenInfo open_info = 1; + public static final int OPEN_INFO_FIELD_NUMBER = 1; private java.util.List openInfo_; public java.util.List getOpenInfoList() { return openInfo_; @@ -4030,7 +4030,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getOpenInfoCount() > 0) { - hash = (37 * hash) + OPENINFO_FIELD_NUMBER; + hash = (37 * hash) + OPEN_INFO_FIELD_NUMBER; hash = (53 * hash) + getOpenInfoList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -4292,7 +4292,7 @@ public final class AdminProtos { private int bitField0_; - // repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1; + // repeated .OpenRegionRequest.RegionOpenInfo open_info = 1; private java.util.List openInfo_ = java.util.Collections.emptyList(); private void ensureOpenInfoIsMutable() { @@ -4492,7 +4492,7 @@ public final class AdminProtos { public interface OpenRegionResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .OpenRegionResponse.RegionOpeningState openingState = 1; + // repeated .OpenRegionResponse.RegionOpeningState opening_state = 1; java.util.List getOpeningStateList(); int getOpeningStateCount(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index); @@ -4597,8 +4597,8 @@ public final class AdminProtos { // @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState) } - // repeated .OpenRegionResponse.RegionOpeningState openingState = 1; - public static final int OPENINGSTATE_FIELD_NUMBER = 1; + // repeated .OpenRegionResponse.RegionOpeningState opening_state = 1; + public static final int OPENING_STATE_FIELD_NUMBER = 1; private java.util.List openingState_; public java.util.List getOpeningStateList() { return openingState_; @@ -4681,7 +4681,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getOpeningStateCount() > 0) { - hash = (37 * hash) + OPENINGSTATE_FIELD_NUMBER; + hash = (37 * hash) + OPENING_STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnumList(getOpeningStateList()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -4931,7 +4931,7 @@ public final class AdminProtos { private int bitField0_; - // repeated .OpenRegionResponse.RegionOpeningState openingState = 1; + // repeated .OpenRegionResponse.RegionOpeningState opening_state = 1; private java.util.List openingState_ = java.util.Collections.emptyList(); private void ensureOpeningStateIsMutable() { @@ -5001,15 +5001,15 @@ public final class AdminProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); - // optional uint32 versionOfClosingNode = 2; + // optional uint32 version_of_closing_node = 2; boolean hasVersionOfClosingNode(); int getVersionOfClosingNode(); - // optional bool transitionInZK = 3 [default = true]; + // optional bool transition_in_ZK = 3 [default = true]; boolean hasTransitionInZK(); boolean getTransitionInZK(); - // optional .ServerName destinationServer = 4; + // optional .ServerName destination_server = 4; boolean hasDestinationServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder(); @@ -5056,8 +5056,8 @@ public final class AdminProtos { return region_; } - // optional uint32 versionOfClosingNode = 2; - public static final int VERSIONOFCLOSINGNODE_FIELD_NUMBER = 2; + // optional uint32 version_of_closing_node = 2; + public static final int VERSION_OF_CLOSING_NODE_FIELD_NUMBER = 2; private int versionOfClosingNode_; public boolean hasVersionOfClosingNode() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -5066,8 +5066,8 @@ public final class AdminProtos { return versionOfClosingNode_; } - // optional bool transitionInZK = 3 [default = true]; - public static final int TRANSITIONINZK_FIELD_NUMBER = 3; + // optional bool transition_in_ZK = 3 [default = true]; + public static final int TRANSITION_IN_ZK_FIELD_NUMBER = 3; private boolean transitionInZK_; public boolean hasTransitionInZK() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -5076,8 +5076,8 @@ public final class AdminProtos { return transitionInZK_; } - // optional .ServerName destinationServer = 4; - public static final int DESTINATIONSERVER_FIELD_NUMBER = 4; + // optional .ServerName destination_server = 4; + public static final int DESTINATION_SERVER_FIELD_NUMBER = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_; public boolean hasDestinationServer() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -5215,15 +5215,15 @@ public final class AdminProtos { hash = (53 * hash) + getRegion().hashCode(); } if (hasVersionOfClosingNode()) { - hash = (37 * hash) + VERSIONOFCLOSINGNODE_FIELD_NUMBER; + hash = (37 * hash) + VERSION_OF_CLOSING_NODE_FIELD_NUMBER; hash = (53 * hash) + getVersionOfClosingNode(); } if (hasTransitionInZK()) { - hash = (37 * hash) + TRANSITIONINZK_FIELD_NUMBER; + hash = (37 * hash) + TRANSITION_IN_ZK_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getTransitionInZK()); } if (hasDestinationServer()) { - hash = (37 * hash) + DESTINATIONSERVER_FIELD_NUMBER; + hash = (37 * hash) + DESTINATION_SERVER_FIELD_NUMBER; hash = (53 * hash) + getDestinationServer().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -5619,7 +5619,7 @@ public final class AdminProtos { return regionBuilder_; } - // optional uint32 versionOfClosingNode = 2; + // optional uint32 version_of_closing_node = 2; private int versionOfClosingNode_ ; public boolean hasVersionOfClosingNode() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -5640,7 +5640,7 @@ public final class AdminProtos { return this; } - // optional bool transitionInZK = 3 [default = true]; + // optional bool transition_in_ZK = 3 [default = true]; private boolean transitionInZK_ = true; public boolean hasTransitionInZK() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -5661,7 +5661,7 @@ public final class AdminProtos { return this; } - // optional .ServerName destinationServer = 4; + // optional .ServerName destination_server = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_; @@ -6149,7 +6149,7 @@ public final class AdminProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); - // optional uint64 ifOlderThanTs = 2; + // optional uint64 if_older_than_ts = 2; boolean hasIfOlderThanTs(); long getIfOlderThanTs(); } @@ -6195,8 +6195,8 @@ public final class AdminProtos { return region_; } - // optional uint64 ifOlderThanTs = 2; - public static final int IFOLDERTHANTS_FIELD_NUMBER = 2; + // optional uint64 if_older_than_ts = 2; + public static final int IF_OLDER_THAN_TS_FIELD_NUMBER = 2; private long ifOlderThanTs_; public boolean hasIfOlderThanTs() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -6299,7 +6299,7 @@ public final class AdminProtos { hash = (53 * hash) + getRegion().hashCode(); } if (hasIfOlderThanTs()) { - hash = (37 * hash) + IFOLDERTHANTS_FIELD_NUMBER; + hash = (37 * hash) + IF_OLDER_THAN_TS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getIfOlderThanTs()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -6648,7 +6648,7 @@ public final class AdminProtos { return regionBuilder_; } - // optional uint64 ifOlderThanTs = 2; + // optional uint64 if_older_than_ts = 2; private long ifOlderThanTs_ ; public boolean hasIfOlderThanTs() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -6683,7 +6683,7 @@ public final class AdminProtos { public interface FlushRegionResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required uint64 lastFlushTime = 1; + // required uint64 last_flush_time = 1; boolean hasLastFlushTime(); long getLastFlushTime(); @@ -6720,8 +6720,8 @@ public final class AdminProtos { } private int bitField0_; - // required uint64 lastFlushTime = 1; - public static final int LASTFLUSHTIME_FIELD_NUMBER = 1; + // required uint64 last_flush_time = 1; + public static final int LAST_FLUSH_TIME_FIELD_NUMBER = 1; private long lastFlushTime_; public boolean hasLastFlushTime() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6826,7 +6826,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLastFlushTime()) { - hash = (37 * hash) + LASTFLUSHTIME_FIELD_NUMBER; + hash = (37 * hash) + LAST_FLUSH_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLastFlushTime()); } if (hasFlushed()) { @@ -7072,7 +7072,7 @@ public final class AdminProtos { private int bitField0_; - // required uint64 lastFlushTime = 1; + // required uint64 last_flush_time = 1; private long lastFlushTime_ ; public boolean hasLastFlushTime() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -7133,7 +7133,7 @@ public final class AdminProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); - // optional bytes splitPoint = 2; + // optional bytes split_point = 2; boolean hasSplitPoint(); com.google.protobuf.ByteString getSplitPoint(); } @@ -7179,8 +7179,8 @@ public final class AdminProtos { return region_; } - // optional bytes splitPoint = 2; - public static final int SPLITPOINT_FIELD_NUMBER = 2; + // optional bytes split_point = 2; + public static final int SPLIT_POINT_FIELD_NUMBER = 2; private com.google.protobuf.ByteString splitPoint_; public boolean hasSplitPoint() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -7283,7 +7283,7 @@ public final class AdminProtos { hash = (53 * hash) + getRegion().hashCode(); } if (hasSplitPoint()) { - hash = (37 * hash) + SPLITPOINT_FIELD_NUMBER; + hash = (37 * hash) + SPLIT_POINT_FIELD_NUMBER; hash = (53 * hash) + getSplitPoint().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -7632,7 +7632,7 @@ public final class AdminProtos { return regionBuilder_; } - // optional bytes splitPoint = 2; + // optional bytes split_point = 2; private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY; public boolean hasSplitPoint() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -8878,12 +8878,12 @@ public final class AdminProtos { public interface MergeRegionsRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .RegionSpecifier regionA = 1; + // required .RegionSpecifier region_a = 1; boolean hasRegionA(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); - // required .RegionSpecifier regionB = 2; + // required .RegionSpecifier region_b = 2; boolean hasRegionB(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); @@ -8921,8 +8921,8 @@ public final class AdminProtos { } private int bitField0_; - // required .RegionSpecifier regionA = 1; - public static final int REGIONA_FIELD_NUMBER = 1; + // required .RegionSpecifier region_a = 1; + public static final int REGION_A_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; public boolean hasRegionA() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -8934,8 +8934,8 @@ public final class AdminProtos { return regionA_; } - // required .RegionSpecifier regionB = 2; - public static final int REGIONB_FIELD_NUMBER = 2; + // required .RegionSpecifier region_b = 2; + public static final int REGION_B_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; public boolean hasRegionB() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -9068,11 +9068,11 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionA()) { - hash = (37 * hash) + REGIONA_FIELD_NUMBER; + hash = (37 * hash) + REGION_A_FIELD_NUMBER; hash = (53 * hash) + getRegionA().hashCode(); } if (hasRegionB()) { - hash = (37 * hash) + REGIONB_FIELD_NUMBER; + hash = (37 * hash) + REGION_B_FIELD_NUMBER; hash = (53 * hash) + getRegionB().hashCode(); } if (hasForcible()) { @@ -9370,7 +9370,7 @@ public final class AdminProtos { private int bitField0_; - // required .RegionSpecifier regionA = 1; + // required .RegionSpecifier region_a = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; @@ -9460,7 +9460,7 @@ public final class AdminProtos { return regionABuilder_; } - // required .RegionSpecifier regionB = 2; + // required .RegionSpecifier region_b = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; @@ -9890,12 +9890,12 @@ public final class AdminProtos { org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey(); org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder(); - // repeated bytes keyValueBytes = 2; + // repeated bytes key_value_bytes = 2; java.util.List getKeyValueBytesList(); int getKeyValueBytesCount(); com.google.protobuf.ByteString getKeyValueBytes(int index); - // optional int32 associatedCellCount = 3; + // optional int32 associated_cell_count = 3; boolean hasAssociatedCellCount(); int getAssociatedCellCount(); } @@ -9941,8 +9941,8 @@ public final class AdminProtos { return key_; } - // repeated bytes keyValueBytes = 2; - public static final int KEYVALUEBYTES_FIELD_NUMBER = 2; + // repeated bytes key_value_bytes = 2; + public static final int KEY_VALUE_BYTES_FIELD_NUMBER = 2; private java.util.List keyValueBytes_; public java.util.List getKeyValueBytesList() { @@ -9955,8 +9955,8 @@ public final class AdminProtos { return keyValueBytes_.get(index); } - // optional int32 associatedCellCount = 3; - public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 3; + // optional int32 associated_cell_count = 3; + public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 3; private int associatedCellCount_; public boolean hasAssociatedCellCount() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -10074,11 +10074,11 @@ public final class AdminProtos { hash = (53 * hash) + getKey().hashCode(); } if (getKeyValueBytesCount() > 0) { - hash = (37 * hash) + KEYVALUEBYTES_FIELD_NUMBER; + hash = (37 * hash) + KEY_VALUE_BYTES_FIELD_NUMBER; hash = (53 * hash) + getKeyValueBytesList().hashCode(); } if (hasAssociatedCellCount()) { - hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER; + hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER; hash = (53 * hash) + getAssociatedCellCount(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -10449,7 +10449,7 @@ public final class AdminProtos { return keyBuilder_; } - // repeated bytes keyValueBytes = 2; + // repeated bytes key_value_bytes = 2; private java.util.List keyValueBytes_ = java.util.Collections.emptyList();; private void ensureKeyValueBytesIsMutable() { if (!((bitField0_ & 0x00000002) == 0x00000002)) { @@ -10500,7 +10500,7 @@ public final class AdminProtos { return this; } - // optional int32 associatedCellCount = 3; + // optional int32 associated_cell_count = 3; private int associatedCellCount_ ; public boolean hasAssociatedCellCount() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -11728,7 +11728,7 @@ public final class AdminProtos { public interface RollWALWriterResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated bytes regionToFlush = 1; + // repeated bytes region_to_flush = 1; java.util.List getRegionToFlushList(); int getRegionToFlushCount(); com.google.protobuf.ByteString getRegionToFlush(int index); @@ -11761,8 +11761,8 @@ public final class AdminProtos { return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable; } - // repeated bytes regionToFlush = 1; - public static final int REGIONTOFLUSH_FIELD_NUMBER = 1; + // repeated bytes region_to_flush = 1; + public static final int REGION_TO_FLUSH_FIELD_NUMBER = 1; private java.util.List regionToFlush_; public java.util.List getRegionToFlushList() { @@ -11846,7 +11846,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getRegionToFlushCount() > 0) { - hash = (37 * hash) + REGIONTOFLUSH_FIELD_NUMBER; + hash = (37 * hash) + REGION_TO_FLUSH_FIELD_NUMBER; hash = (53 * hash) + getRegionToFlushList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -12076,7 +12076,7 @@ public final class AdminProtos { private int bitField0_; - // repeated bytes regionToFlush = 1; + // repeated bytes region_to_flush = 1; private java.util.List regionToFlush_ = java.util.Collections.emptyList();; private void ensureRegionToFlushIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { @@ -13157,12 +13157,12 @@ public final class AdminProtos { public interface ServerInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .ServerName serverName = 1; + // required .ServerName server_name = 1; boolean hasServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); - // optional uint32 webuiPort = 2; + // optional uint32 webui_port = 2; boolean hasWebuiPort(); int getWebuiPort(); } @@ -13195,8 +13195,8 @@ public final class AdminProtos { } private int bitField0_; - // required .ServerName serverName = 1; - public static final int SERVERNAME_FIELD_NUMBER = 1; + // required .ServerName server_name = 1; + public static final int SERVER_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; public boolean hasServerName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -13208,8 +13208,8 @@ public final class AdminProtos { return serverName_; } - // optional uint32 webuiPort = 2; - public static final int WEBUIPORT_FIELD_NUMBER = 2; + // optional uint32 webui_port = 2; + public static final int WEBUI_PORT_FIELD_NUMBER = 2; private int webuiPort_; public boolean hasWebuiPort() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -13308,11 +13308,11 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasServerName()) { - hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; + hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } if (hasWebuiPort()) { - hash = (37 * hash) + WEBUIPORT_FIELD_NUMBER; + hash = (37 * hash) + WEBUI_PORT_FIELD_NUMBER; hash = (53 * hash) + getWebuiPort(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -13571,7 +13571,7 @@ public final class AdminProtos { private int bitField0_; - // required .ServerName serverName = 1; + // required .ServerName server_name = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; @@ -13661,7 +13661,7 @@ public final class AdminProtos { return serverNameBuilder_; } - // optional uint32 webuiPort = 2; + // optional uint32 webui_port = 2; private int webuiPort_ ; public boolean hasWebuiPort() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -13696,7 +13696,7 @@ public final class AdminProtos { public interface GetServerInfoResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .ServerInfo serverInfo = 1; + // required .ServerInfo server_info = 1; boolean hasServerInfo(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder(); @@ -13730,8 +13730,8 @@ public final class AdminProtos { } private int bitField0_; - // required .ServerInfo serverInfo = 1; - public static final int SERVERINFO_FIELD_NUMBER = 1; + // required .ServerInfo server_info = 1; + public static final int SERVER_INFO_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_; public boolean hasServerInfo() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -13820,7 +13820,7 @@ public final class AdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasServerInfo()) { - hash = (37 * hash) + SERVERINFO_FIELD_NUMBER; + hash = (37 * hash) + SERVER_INFO_FIELD_NUMBER; hash = (53 * hash) + getServerInfo().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -14065,7 +14065,7 @@ public final class AdminProtos { private int bitField0_; - // required .ServerInfo serverInfo = 1; + // required .ServerInfo server_info = 1; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_; @@ -15387,79 +15387,80 @@ public final class AdminProtos { static { java.lang.String[] descriptorData = { "\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" + - "\032\tWAL.proto\"Q\n\024GetRegionInfoRequest\022 \n\006r" + - "egion\030\001 \002(\0132\020.RegionSpecifier\022\027\n\017compact" + - "ionState\030\002 \001(\010\"\301\001\n\025GetRegionInfoResponse" + - "\022\037\n\nregionInfo\030\001 \002(\0132\013.RegionInfo\022?\n\017com" + - "pactionState\030\002 \001(\0162&.GetRegionInfoRespon" + - "se.CompactionState\"F\n\017CompactionState\022\010\n" + - "\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJOR_A" + - "ND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006reg" + - "ion\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family\030\002 ", - "\003(\014\")\n\024GetStoreFileResponse\022\021\n\tstoreFile" + - "\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\":\n\027GetO" + - "nlineRegionResponse\022\037\n\nregionInfo\030\001 \003(\0132" + - "\013.RegionInfo\"\270\001\n\021OpenRegionRequest\0223\n\010op" + - "enInfo\030\001 \003(\0132!.OpenRegionRequest.RegionO" + - "penInfo\032n\n\016RegionOpenInfo\022\033\n\006region\030\001 \002(" + - "\0132\013.RegionInfo\022\034\n\024versionOfOfflineNode\030\002" + - " \001(\r\022!\n\014favoredNodes\030\003 \003(\0132\013.ServerName\"" + - "\234\001\n\022OpenRegionResponse\022<\n\014openingState\030\001" + - " \003(\0162&.OpenRegionResponse.RegionOpeningS", - "tate\"H\n\022RegionOpeningState\022\n\n\006OPENED\020\000\022\022" + - "\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_OPENING\020\002\"\232" + - "\001\n\022CloseRegionRequest\022 \n\006region\030\001 \002(\0132\020." + - "RegionSpecifier\022\034\n\024versionOfClosingNode\030" + - "\002 \001(\r\022\034\n\016transitionInZK\030\003 \001(\010:\004true\022&\n\021d" + - "estinationServer\030\004 \001(\0132\013.ServerName\"%\n\023C" + - "loseRegionResponse\022\016\n\006closed\030\001 \002(\010\"M\n\022Fl" + - "ushRegionRequest\022 \n\006region\030\001 \002(\0132\020.Regio" + - "nSpecifier\022\025\n\rifOlderThanTs\030\002 \001(\004\"=\n\023Flu" + - "shRegionResponse\022\025\n\rlastFlushTime\030\001 \002(\004\022", - "\017\n\007flushed\030\002 \001(\010\"J\n\022SplitRegionRequest\022 " + - "\n\006region\030\001 \002(\0132\020.RegionSpecifier\022\022\n\nspli" + - "tPoint\030\002 \001(\014\"\025\n\023SplitRegionResponse\"W\n\024C" + - "ompactRegionRequest\022 \n\006region\030\001 \002(\0132\020.Re" + - "gionSpecifier\022\r\n\005major\030\002 \001(\010\022\016\n\006family\030\003" + - " \001(\014\"\027\n\025CompactRegionResponse\"t\n\023MergeRe" + - "gionsRequest\022!\n\007regionA\030\001 \002(\0132\020.RegionSp" + - "ecifier\022!\n\007regionB\030\002 \002(\0132\020.RegionSpecifi" + - "er\022\027\n\010forcible\030\003 \001(\010:\005false\"\026\n\024MergeRegi" + - "onsResponse\"T\n\010WALEntry\022\024\n\003key\030\001 \002(\0132\007.W", - "ALKey\022\025\n\rkeyValueBytes\030\002 \003(\014\022\033\n\023associat" + - "edCellCount\030\003 \001(\005\"4\n\030ReplicateWALEntryRe" + - "quest\022\030\n\005entry\030\001 \003(\0132\t.WALEntry\"\033\n\031Repli" + - "cateWALEntryResponse\"\026\n\024RollWALWriterReq" + - "uest\".\n\025RollWALWriterResponse\022\025\n\rregionT" + - "oFlush\030\001 \003(\014\"#\n\021StopServerRequest\022\016\n\006rea" + - "son\030\001 \002(\t\"\024\n\022StopServerResponse\"\026\n\024GetSe" + - "rverInfoRequest\"@\n\nServerInfo\022\037\n\nserverN" + - "ame\030\001 \002(\0132\013.ServerName\022\021\n\twebuiPort\030\002 \001(" + - "\r\"8\n\025GetServerInfoResponse\022\037\n\nserverInfo", - "\030\001 \002(\0132\013.ServerInfo2\337\006\n\014AdminService\022>\n\r" + - "getRegionInfo\022\025.GetRegionInfoRequest\032\026.G" + - "etRegionInfoResponse\022;\n\014getStoreFile\022\024.G" + - "etStoreFileRequest\032\025.GetStoreFileRespons" + - "e\022D\n\017getOnlineRegion\022\027.GetOnlineRegionRe" + - "quest\032\030.GetOnlineRegionResponse\0225\n\nopenR" + - "egion\022\022.OpenRegionRequest\032\023.OpenRegionRe" + - "sponse\0228\n\013closeRegion\022\023.CloseRegionReque" + - "st\032\024.CloseRegionResponse\0228\n\013flushRegion\022" + - "\023.FlushRegionRequest\032\024.FlushRegionRespon", - "se\0228\n\013splitRegion\022\023.SplitRegionRequest\032\024" + - ".SplitRegionResponse\022>\n\rcompactRegion\022\025." + - "CompactRegionRequest\032\026.CompactRegionResp" + - "onse\022;\n\014mergeRegions\022\024.MergeRegionsReque" + - "st\032\025.MergeRegionsResponse\022J\n\021replicateWA" + - "LEntry\022\031.ReplicateWALEntryRequest\032\032.Repl" + - "icateWALEntryResponse\022\'\n\006replay\022\r.MultiR" + - "equest\032\016.MultiResponse\022>\n\rrollWALWriter\022" + - "\025.RollWALWriterRequest\032\026.RollWALWriterRe" + - "sponse\022>\n\rgetServerInfo\022\025.GetServerInfoR", - "equest\032\026.GetServerInfoResponse\0225\n\nstopSe" + - "rver\022\022.StopServerRequest\032\023.StopServerRes" + - "ponseBA\n*org.apache.hadoop.hbase.protobu" + - "f.generatedB\013AdminProtosH\001\210\001\001\240\001\001" + "\032\tWAL.proto\"R\n\024GetRegionInfoRequest\022 \n\006r" + + "egion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020compact" + + "ion_state\030\002 \001(\010\"\303\001\n\025GetRegionInfoRespons" + + "e\022 \n\013region_info\030\001 \002(\0132\013.RegionInfo\022@\n\020c" + + "ompaction_state\030\002 \001(\0162&.GetRegionInfoRes" + + "ponse.CompactionState\"F\n\017CompactionState" + + "\022\010\n\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJO" + + "R_AND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006" + + "region\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family", + "\030\002 \003(\014\"*\n\024GetStoreFileResponse\022\022\n\nstore_" + + "file\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\";\n\027" + + "GetOnlineRegionResponse\022 \n\013region_info\030\001" + + " \003(\0132\013.RegionInfo\"\275\001\n\021OpenRegionRequest\022" + + "4\n\topen_info\030\001 \003(\0132!.OpenRegionRequest.R" + + "egionOpenInfo\032r\n\016RegionOpenInfo\022\033\n\006regio" + + "n\030\001 \002(\0132\013.RegionInfo\022\037\n\027version_of_offli" + + "ne_node\030\002 \001(\r\022\"\n\rfavored_nodes\030\003 \003(\0132\013.S" + + "erverName\"\235\001\n\022OpenRegionResponse\022=\n\ropen" + + "ing_state\030\001 \003(\0162&.OpenRegionResponse.Reg", + "ionOpeningState\"H\n\022RegionOpeningState\022\n\n" + + "\006OPENED\020\000\022\022\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_" + + "OPENING\020\002\"\240\001\n\022CloseRegionRequest\022 \n\006regi" + + "on\030\001 \002(\0132\020.RegionSpecifier\022\037\n\027version_of" + + "_closing_node\030\002 \001(\r\022\036\n\020transition_in_ZK\030" + + "\003 \001(\010:\004true\022\'\n\022destination_server\030\004 \001(\0132" + + "\013.ServerName\"%\n\023CloseRegionResponse\022\016\n\006c" + + "losed\030\001 \002(\010\"P\n\022FlushRegionRequest\022 \n\006reg" + + "ion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020if_older_" + + "than_ts\030\002 \001(\004\"?\n\023FlushRegionResponse\022\027\n\017", + "last_flush_time\030\001 \002(\004\022\017\n\007flushed\030\002 \001(\010\"K" + + "\n\022SplitRegionRequest\022 \n\006region\030\001 \002(\0132\020.R" + + "egionSpecifier\022\023\n\013split_point\030\002 \001(\014\"\025\n\023S" + + "plitRegionResponse\"W\n\024CompactRegionReque" + + "st\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\022\r\n\005" + + "major\030\002 \001(\010\022\016\n\006family\030\003 \001(\014\"\027\n\025CompactRe" + + "gionResponse\"v\n\023MergeRegionsRequest\022\"\n\010r" + + "egion_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010regio" + + "n_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcible\030" + + "\003 \001(\010:\005false\"\026\n\024MergeRegionsResponse\"X\n\010", + "WALEntry\022\024\n\003key\030\001 \002(\0132\007.WALKey\022\027\n\017key_va" + + "lue_bytes\030\002 \003(\014\022\035\n\025associated_cell_count" + + "\030\003 \001(\005\"4\n\030ReplicateWALEntryRequest\022\030\n\005en" + + "try\030\001 \003(\0132\t.WALEntry\"\033\n\031ReplicateWALEntr" + + "yResponse\"\026\n\024RollWALWriterRequest\"0\n\025Rol" + + "lWALWriterResponse\022\027\n\017region_to_flush\030\001 " + + "\003(\014\"#\n\021StopServerRequest\022\016\n\006reason\030\001 \002(\t" + + "\"\024\n\022StopServerResponse\"\026\n\024GetServerInfoR" + + "equest\"B\n\nServerInfo\022 \n\013server_name\030\001 \002(" + + "\0132\013.ServerName\022\022\n\nwebui_port\030\002 \001(\r\"9\n\025Ge", + "tServerInfoResponse\022 \n\013server_info\030\001 \002(\013" + + "2\013.ServerInfo2\337\006\n\014AdminService\022>\n\rGetReg" + + "ionInfo\022\025.GetRegionInfoRequest\032\026.GetRegi" + + "onInfoResponse\022;\n\014GetStoreFile\022\024.GetStor" + + "eFileRequest\032\025.GetStoreFileResponse\022D\n\017G" + + "etOnlineRegion\022\027.GetOnlineRegionRequest\032" + + "\030.GetOnlineRegionResponse\0225\n\nOpenRegion\022" + + "\022.OpenRegionRequest\032\023.OpenRegionResponse" + + "\0228\n\013CloseRegion\022\023.CloseRegionRequest\032\024.C" + + "loseRegionResponse\0228\n\013FlushRegion\022\023.Flus", + "hRegionRequest\032\024.FlushRegionResponse\0228\n\013" + + "SplitRegion\022\023.SplitRegionRequest\032\024.Split" + + "RegionResponse\022>\n\rCompactRegion\022\025.Compac" + + "tRegionRequest\032\026.CompactRegionResponse\022;" + + "\n\014MergeRegions\022\024.MergeRegionsRequest\032\025.M" + + "ergeRegionsResponse\022J\n\021ReplicateWALEntry" + + "\022\031.ReplicateWALEntryRequest\032\032.ReplicateW" + + "ALEntryResponse\022\'\n\006Replay\022\r.MultiRequest" + + "\032\016.MultiResponse\022>\n\rRollWALWriter\022\025.Roll" + + "WALWriterRequest\032\026.RollWALWriterResponse", + "\022>\n\rGetServerInfo\022\025.GetServerInfoRequest" + + "\032\026.GetServerInfoResponse\0225\n\nStopServer\022\022" + + ".StopServerRequest\032\023.StopServerResponseB" + + "A\n*org.apache.hadoop.hbase.protobuf.gene" + + "ratedB\013AdminProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java index ad1dbeb..bbfd3fd 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java @@ -11,7 +11,7 @@ public final class AggregateProtos { public interface AggregateArgumentOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string interpreterClassName = 1; + // required string interpreter_class_name = 1; boolean hasInterpreterClassName(); String getInterpreterClassName(); @@ -20,7 +20,7 @@ public final class AggregateProtos { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder(); - // optional bytes interpreterSpecificBytes = 3; + // optional bytes interpreter_specific_bytes = 3; boolean hasInterpreterSpecificBytes(); com.google.protobuf.ByteString getInterpreterSpecificBytes(); } @@ -53,8 +53,8 @@ public final class AggregateProtos { } private int bitField0_; - // required string interpreterClassName = 1; - public static final int INTERPRETERCLASSNAME_FIELD_NUMBER = 1; + // required string interpreter_class_name = 1; + public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1; private java.lang.Object interpreterClassName_; public boolean hasInterpreterClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -98,8 +98,8 @@ public final class AggregateProtos { return scan_; } - // optional bytes interpreterSpecificBytes = 3; - public static final int INTERPRETERSPECIFICBYTES_FIELD_NUMBER = 3; + // optional bytes interpreter_specific_bytes = 3; + public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3; private com.google.protobuf.ByteString interpreterSpecificBytes_; public boolean hasInterpreterSpecificBytes() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -215,7 +215,7 @@ public final class AggregateProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasInterpreterClassName()) { - hash = (37 * hash) + INTERPRETERCLASSNAME_FIELD_NUMBER; + hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getInterpreterClassName().hashCode(); } if (hasScan()) { @@ -223,7 +223,7 @@ public final class AggregateProtos { hash = (53 * hash) + getScan().hashCode(); } if (hasInterpreterSpecificBytes()) { - hash = (37 * hash) + INTERPRETERSPECIFICBYTES_FIELD_NUMBER; + hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER; hash = (53 * hash) + getInterpreterSpecificBytes().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -500,7 +500,7 @@ public final class AggregateProtos { private int bitField0_; - // required string interpreterClassName = 1; + // required string interpreter_class_name = 1; private java.lang.Object interpreterClassName_ = ""; public boolean hasInterpreterClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -626,7 +626,7 @@ public final class AggregateProtos { return scanBuilder_; } - // optional bytes interpreterSpecificBytes = 3; + // optional bytes interpreter_specific_bytes = 3; private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY; public boolean hasInterpreterSpecificBytes() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -664,12 +664,12 @@ public final class AggregateProtos { public interface AggregateResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated bytes firstPart = 1; + // repeated bytes first_part = 1; java.util.List getFirstPartList(); int getFirstPartCount(); com.google.protobuf.ByteString getFirstPart(int index); - // optional bytes secondPart = 2; + // optional bytes second_part = 2; boolean hasSecondPart(); com.google.protobuf.ByteString getSecondPart(); } @@ -702,8 +702,8 @@ public final class AggregateProtos { } private int bitField0_; - // repeated bytes firstPart = 1; - public static final int FIRSTPART_FIELD_NUMBER = 1; + // repeated bytes first_part = 1; + public static final int FIRST_PART_FIELD_NUMBER = 1; private java.util.List firstPart_; public java.util.List getFirstPartList() { @@ -716,8 +716,8 @@ public final class AggregateProtos { return firstPart_.get(index); } - // optional bytes secondPart = 2; - public static final int SECONDPART_FIELD_NUMBER = 2; + // optional bytes second_part = 2; + public static final int SECOND_PART_FIELD_NUMBER = 2; private com.google.protobuf.ByteString secondPart_; public boolean hasSecondPart() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -810,11 +810,11 @@ public final class AggregateProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getFirstPartCount() > 0) { - hash = (37 * hash) + FIRSTPART_FIELD_NUMBER; + hash = (37 * hash) + FIRST_PART_FIELD_NUMBER; hash = (53 * hash) + getFirstPartList().hashCode(); } if (hasSecondPart()) { - hash = (37 * hash) + SECONDPART_FIELD_NUMBER; + hash = (37 * hash) + SECOND_PART_FIELD_NUMBER; hash = (53 * hash) + getSecondPart().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1060,7 +1060,7 @@ public final class AggregateProtos { private int bitField0_; - // repeated bytes firstPart = 1; + // repeated bytes first_part = 1; private java.util.List firstPart_ = java.util.Collections.emptyList();; private void ensureFirstPartIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1111,7 +1111,7 @@ public final class AggregateProtos { return this; } - // optional bytes secondPart = 2; + // optional bytes second_part = 2; private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY; public boolean hasSecondPart() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1776,23 +1776,23 @@ public final class AggregateProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\017Aggregate.proto\032\014Client.proto\"h\n\021Aggre" + - "gateArgument\022\034\n\024interpreterClassName\030\001 \002" + - "(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022 \n\030interpreterSp" + - "ecificBytes\030\003 \001(\014\":\n\021AggregateResponse\022\021" + - "\n\tfirstPart\030\001 \003(\014\022\022\n\nsecondPart\030\002 \001(\0142\366\002" + - "\n\020AggregateService\0220\n\006getMax\022\022.Aggregate" + - "Argument\032\022.AggregateResponse\0220\n\006getMin\022\022" + - ".AggregateArgument\032\022.AggregateResponse\0220" + - "\n\006getSum\022\022.AggregateArgument\032\022.Aggregate" + - "Response\0223\n\tgetRowNum\022\022.AggregateArgumen", - "t\032\022.AggregateResponse\0220\n\006getAvg\022\022.Aggreg" + - "ateArgument\032\022.AggregateResponse\0220\n\006getSt" + - "d\022\022.AggregateArgument\032\022.AggregateRespons" + - "e\0223\n\tgetMedian\022\022.AggregateArgument\032\022.Agg" + - "regateResponseBE\n*org.apache.hadoop.hbas" + - "e.protobuf.generatedB\017AggregateProtosH\001\210" + - "\001\001\240\001\001" + "\n\017Aggregate.proto\032\014Client.proto\"l\n\021Aggre" + + "gateArgument\022\036\n\026interpreter_class_name\030\001" + + " \002(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022\"\n\032interpreter" + + "_specific_bytes\030\003 \001(\014\"<\n\021AggregateRespon" + + "se\022\022\n\nfirst_part\030\001 \003(\014\022\023\n\013second_part\030\002 " + + "\001(\0142\366\002\n\020AggregateService\0220\n\006GetMax\022\022.Agg" + + "regateArgument\032\022.AggregateResponse\0220\n\006Ge" + + "tMin\022\022.AggregateArgument\032\022.AggregateResp" + + "onse\0220\n\006GetSum\022\022.AggregateArgument\032\022.Agg" + + "regateResponse\0223\n\tGetRowNum\022\022.AggregateA", + "rgument\032\022.AggregateResponse\0220\n\006GetAvg\022\022." + + "AggregateArgument\032\022.AggregateResponse\0220\n" + + "\006GetStd\022\022.AggregateArgument\032\022.AggregateR" + + "esponse\0223\n\tGetMedian\022\022.AggregateArgument" + + "\032\022.AggregateResponseBE\n*org.apache.hadoo" + + "p.hbase.protobuf.generatedB\017AggregatePro" + + "tosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AuthenticationProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AuthenticationProtos.java index 5a05473..df0be70 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AuthenticationProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AuthenticationProtos.java @@ -15,7 +15,7 @@ public final class AuthenticationProtos { boolean hasId(); int getId(); - // required int64 expirationDate = 2; + // required int64 expiration_date = 2; boolean hasExpirationDate(); long getExpirationDate(); @@ -62,8 +62,8 @@ public final class AuthenticationProtos { return id_; } - // required int64 expirationDate = 2; - public static final int EXPIRATIONDATE_FIELD_NUMBER = 2; + // required int64 expiration_date = 2; + public static final int EXPIRATION_DATE_FIELD_NUMBER = 2; private long expirationDate_; public boolean hasExpirationDate() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -193,7 +193,7 @@ public final class AuthenticationProtos { hash = (53 * hash) + getId(); } if (hasExpirationDate()) { - hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER; + hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getExpirationDate()); } if (hasKey()) { @@ -482,7 +482,7 @@ public final class AuthenticationProtos { return this; } - // required int64 expirationDate = 2; + // required int64 expiration_date = 2; private long expirationDate_ ; public boolean hasExpirationDate() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -549,19 +549,19 @@ public final class AuthenticationProtos { boolean hasUsername(); com.google.protobuf.ByteString getUsername(); - // required int32 keyId = 3; + // required int32 key_id = 3; boolean hasKeyId(); int getKeyId(); - // optional int64 issueDate = 4; + // optional int64 issue_date = 4; boolean hasIssueDate(); long getIssueDate(); - // optional int64 expirationDate = 5; + // optional int64 expiration_date = 5; boolean hasExpirationDate(); long getExpirationDate(); - // optional int64 sequenceNumber = 6; + // optional int64 sequence_number = 6; boolean hasSequenceNumber(); long getSequenceNumber(); } @@ -680,8 +680,8 @@ public final class AuthenticationProtos { return username_; } - // required int32 keyId = 3; - public static final int KEYID_FIELD_NUMBER = 3; + // required int32 key_id = 3; + public static final int KEY_ID_FIELD_NUMBER = 3; private int keyId_; public boolean hasKeyId() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -690,8 +690,8 @@ public final class AuthenticationProtos { return keyId_; } - // optional int64 issueDate = 4; - public static final int ISSUEDATE_FIELD_NUMBER = 4; + // optional int64 issue_date = 4; + public static final int ISSUE_DATE_FIELD_NUMBER = 4; private long issueDate_; public boolean hasIssueDate() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -700,8 +700,8 @@ public final class AuthenticationProtos { return issueDate_; } - // optional int64 expirationDate = 5; - public static final int EXPIRATIONDATE_FIELD_NUMBER = 5; + // optional int64 expiration_date = 5; + public static final int EXPIRATION_DATE_FIELD_NUMBER = 5; private long expirationDate_; public boolean hasExpirationDate() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -710,8 +710,8 @@ public final class AuthenticationProtos { return expirationDate_; } - // optional int64 sequenceNumber = 6; - public static final int SEQUENCENUMBER_FIELD_NUMBER = 6; + // optional int64 sequence_number = 6; + public static final int SEQUENCE_NUMBER_FIELD_NUMBER = 6; private long sequenceNumber_; public boolean hasSequenceNumber() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -874,19 +874,19 @@ public final class AuthenticationProtos { hash = (53 * hash) + getUsername().hashCode(); } if (hasKeyId()) { - hash = (37 * hash) + KEYID_FIELD_NUMBER; + hash = (37 * hash) + KEY_ID_FIELD_NUMBER; hash = (53 * hash) + getKeyId(); } if (hasIssueDate()) { - hash = (37 * hash) + ISSUEDATE_FIELD_NUMBER; + hash = (37 * hash) + ISSUE_DATE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getIssueDate()); } if (hasExpirationDate()) { - hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER; + hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getExpirationDate()); } if (hasSequenceNumber()) { - hash = (37 * hash) + SEQUENCENUMBER_FIELD_NUMBER; + hash = (37 * hash) + SEQUENCE_NUMBER_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSequenceNumber()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1246,7 +1246,7 @@ public final class AuthenticationProtos { return this; } - // required int32 keyId = 3; + // required int32 key_id = 3; private int keyId_ ; public boolean hasKeyId() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1267,7 +1267,7 @@ public final class AuthenticationProtos { return this; } - // optional int64 issueDate = 4; + // optional int64 issue_date = 4; private long issueDate_ ; public boolean hasIssueDate() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1288,7 +1288,7 @@ public final class AuthenticationProtos { return this; } - // optional int64 expirationDate = 5; + // optional int64 expiration_date = 5; private long expirationDate_ ; public boolean hasExpirationDate() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -1309,7 +1309,7 @@ public final class AuthenticationProtos { return this; } - // optional int64 sequenceNumber = 6; + // optional int64 sequence_number = 6; private long sequenceNumber_ ; public boolean hasSequenceNumber() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -2917,7 +2917,7 @@ public final class AuthenticationProtos { boolean hasUsername(); String getUsername(); - // optional string authMethod = 2; + // optional string auth_method = 2; boolean hasAuthMethod(); String getAuthMethod(); } @@ -2982,8 +2982,8 @@ public final class AuthenticationProtos { } } - // optional string authMethod = 2; - public static final int AUTHMETHOD_FIELD_NUMBER = 2; + // optional string auth_method = 2; + public static final int AUTH_METHOD_FIELD_NUMBER = 2; private java.lang.Object authMethod_; public boolean hasAuthMethod() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3100,7 +3100,7 @@ public final class AuthenticationProtos { hash = (53 * hash) + getUsername().hashCode(); } if (hasAuthMethod()) { - hash = (37 * hash) + AUTHMETHOD_FIELD_NUMBER; + hash = (37 * hash) + AUTH_METHOD_FIELD_NUMBER; hash = (53 * hash) + getAuthMethod().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3374,7 +3374,7 @@ public final class AuthenticationProtos { onChanged(); } - // optional string authMethod = 2; + // optional string auth_method = 2; private java.lang.Object authMethod_ = ""; public boolean hasAuthMethod() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3431,7 +3431,7 @@ public final class AuthenticationProtos { org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request, com.google.protobuf.RpcCallback done); - public abstract void whoami( + public abstract void whoAmI( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, com.google.protobuf.RpcCallback done); @@ -3450,11 +3450,11 @@ public final class AuthenticationProtos { } @java.lang.Override - public void whoami( + public void whoAmI( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, com.google.protobuf.RpcCallback done) { - impl.whoami(controller, request, done); + impl.whoAmI(controller, request, done); } }; @@ -3482,7 +3482,7 @@ public final class AuthenticationProtos { case 0: return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request); case 1: - return impl.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request); + return impl.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request); default: throw new java.lang.AssertionError("Can't get here."); } @@ -3532,7 +3532,7 @@ public final class AuthenticationProtos { org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request, com.google.protobuf.RpcCallback done); - public abstract void whoami( + public abstract void whoAmI( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, com.google.protobuf.RpcCallback done); @@ -3565,7 +3565,7 @@ public final class AuthenticationProtos { done)); return; case 1: - this.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request, + this.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request, com.google.protobuf.RpcUtil.specializeCallback( done)); return; @@ -3641,7 +3641,7 @@ public final class AuthenticationProtos { org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance())); } - public void whoami( + public void whoAmI( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, com.google.protobuf.RpcCallback done) { @@ -3668,7 +3668,7 @@ public final class AuthenticationProtos { org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request) throws com.google.protobuf.ServiceException; - public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami( + public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request) throws com.google.protobuf.ServiceException; @@ -3693,7 +3693,7 @@ public final class AuthenticationProtos { } - public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami( + public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request) throws com.google.protobuf.ServiceException { @@ -3751,23 +3751,23 @@ public final class AuthenticationProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\024Authentication.proto\"D\n\021Authentication" + - "Key\022\n\n\002id\030\001 \002(\005\022\026\n\016expirationDate\030\002 \002(\003\022" + - "\013\n\003key\030\003 \002(\014\"\270\001\n\017TokenIdentifier\022#\n\004kind" + - "\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010usernam" + - "e\030\002 \002(\014\022\r\n\005keyId\030\003 \002(\005\022\021\n\tissueDate\030\004 \001(" + - "\003\022\026\n\016expirationDate\030\005 \001(\003\022\026\n\016sequenceNum" + - "ber\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOKEN\020\000\"" + - ">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" + - "\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenRequest\"" + - "&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.Token\"", - "\017\n\rWhoAmIRequest\"6\n\016WhoAmIResponse\022\020\n\010us" + - "ername\030\001 \001(\t\022\022\n\nauthMethod\030\002 \001(\t2{\n\025Auth" + - "enticationService\0227\n\026getAuthenticationTo" + - "ken\022\r.TokenRequest\032\016.TokenResponse\022)\n\006wh" + - "oami\022\016.WhoAmIRequest\032\017.WhoAmIResponseBJ\n" + - "*org.apache.hadoop.hbase.protobuf.genera" + - "tedB\024AuthenticationProtosH\001\210\001\001\240\001\001" + "\n\024Authentication.proto\"E\n\021Authentication" + + "Key\022\n\n\002id\030\001 \002(\005\022\027\n\017expiration_date\030\002 \002(\003" + + "\022\013\n\003key\030\003 \002(\014\"\274\001\n\017TokenIdentifier\022#\n\004kin" + + "d\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010userna" + + "me\030\002 \002(\014\022\016\n\006key_id\030\003 \002(\005\022\022\n\nissue_date\030\004" + + " \001(\003\022\027\n\017expiration_date\030\005 \001(\003\022\027\n\017sequenc" + + "e_number\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOK" + + "EN\020\000\">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" + + "sword\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenReq" + + "uest\"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.T", + "oken\"\017\n\rWhoAmIRequest\"7\n\016WhoAmIResponse\022" + + "\020\n\010username\030\001 \001(\t\022\023\n\013auth_method\030\002 \001(\t2{" + + "\n\025AuthenticationService\0227\n\026GetAuthentica" + + "tionToken\022\r.TokenRequest\032\016.TokenResponse" + + "\022)\n\006WhoAmI\022\016.WhoAmIRequest\032\017.WhoAmIRespo" + + "nseBJ\n*org.apache.hadoop.hbase.protobuf." + + "generatedB\024AuthenticationProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java index 4408081..228d63e 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/CellProtos.java @@ -108,7 +108,7 @@ public final class CellProtos { boolean hasTimestamp(); long getTimestamp(); - // optional .CellType cellType = 5; + // optional .CellType cell_type = 5; boolean hasCellType(); org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType(); @@ -185,8 +185,8 @@ public final class CellProtos { return timestamp_; } - // optional .CellType cellType = 5; - public static final int CELLTYPE_FIELD_NUMBER = 5; + // optional .CellType cell_type = 5; + public static final int CELL_TYPE_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_; public boolean hasCellType() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -355,7 +355,7 @@ public final class CellProtos { hash = (53 * hash) + hashLong(getTimestamp()); } if (hasCellType()) { - hash = (37 * hash) + CELLTYPE_FIELD_NUMBER; + hash = (37 * hash) + CELL_TYPE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getCellType()); } if (hasValue()) { @@ -752,7 +752,7 @@ public final class CellProtos { return this; } - // optional .CellType cellType = 5; + // optional .CellType cell_type = 5; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; public boolean hasCellType() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -830,7 +830,7 @@ public final class CellProtos { boolean hasTimestamp(); long getTimestamp(); - // optional .CellType keyType = 5; + // optional .CellType key_type = 5; boolean hasKeyType(); org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType(); @@ -907,8 +907,8 @@ public final class CellProtos { return timestamp_; } - // optional .CellType keyType = 5; - public static final int KEYTYPE_FIELD_NUMBER = 5; + // optional .CellType key_type = 5; + public static final int KEY_TYPE_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_; public boolean hasKeyType() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -1089,7 +1089,7 @@ public final class CellProtos { hash = (53 * hash) + hashLong(getTimestamp()); } if (hasKeyType()) { - hash = (37 * hash) + KEYTYPE_FIELD_NUMBER; + hash = (37 * hash) + KEY_TYPE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getKeyType()); } if (hasValue()) { @@ -1498,7 +1498,7 @@ public final class CellProtos { return this; } - // optional .CellType keyType = 5; + // optional .CellType key_type = 5; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; public boolean hasKeyType() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -1576,17 +1576,17 @@ public final class CellProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\nCell.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" + + "\n\nCell.proto\"v\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" + "ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" + - "p\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n\005" + - "value\030\006 \001(\014\"x\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n\006" + - "family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttimes" + - "tamp\030\004 \001(\004\022\032\n\007keyType\030\005 \001(\0162\t.CellType\022\r" + - "\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000\022\007" + - "\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014\022\021" + - "\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org.a" + - "pache.hadoop.hbase.protobuf.generatedB\nC", - "ellProtosH\001\240\001\001" + "p\030\004 \001(\004\022\034\n\tcell_type\030\005 \001(\0162\t.CellType\022\r\n" + + "\005value\030\006 \001(\014\"y\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n" + + "\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttime" + + "stamp\030\004 \001(\004\022\033\n\010key_type\030\005 \001(\0162\t.CellType" + + "\022\r\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000" + + "\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014" + + "\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org" + + ".apache.hadoop.hbase.protobuf.generatedB", + "\nCellProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java index b996e82..6d8dce7 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java @@ -533,24 +533,24 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); - // optional .TimeRange timeRange = 5; + // optional .TimeRange time_range = 5; boolean hasTimeRange(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange getTimeRange(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder getTimeRangeOrBuilder(); - // optional uint32 maxVersions = 6 [default = 1]; + // optional uint32 max_versions = 6 [default = 1]; boolean hasMaxVersions(); int getMaxVersions(); - // optional bool cacheBlocks = 7 [default = true]; + // optional bool cache_blocks = 7 [default = true]; boolean hasCacheBlocks(); boolean getCacheBlocks(); - // optional uint32 storeLimit = 8; + // optional uint32 store_limit = 8; boolean hasStoreLimit(); int getStoreLimit(); - // optional uint32 storeOffset = 9; + // optional uint32 store_offset = 9; boolean hasStoreOffset(); int getStoreOffset(); } @@ -648,8 +648,8 @@ public final class ClientProtos { return filter_; } - // optional .TimeRange timeRange = 5; - public static final int TIMERANGE_FIELD_NUMBER = 5; + // optional .TimeRange time_range = 5; + public static final int TIME_RANGE_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange timeRange_; public boolean hasTimeRange() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -661,8 +661,8 @@ public final class ClientProtos { return timeRange_; } - // optional uint32 maxVersions = 6 [default = 1]; - public static final int MAXVERSIONS_FIELD_NUMBER = 6; + // optional uint32 max_versions = 6 [default = 1]; + public static final int MAX_VERSIONS_FIELD_NUMBER = 6; private int maxVersions_; public boolean hasMaxVersions() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -671,8 +671,8 @@ public final class ClientProtos { return maxVersions_; } - // optional bool cacheBlocks = 7 [default = true]; - public static final int CACHEBLOCKS_FIELD_NUMBER = 7; + // optional bool cache_blocks = 7 [default = true]; + public static final int CACHE_BLOCKS_FIELD_NUMBER = 7; private boolean cacheBlocks_; public boolean hasCacheBlocks() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -681,8 +681,8 @@ public final class ClientProtos { return cacheBlocks_; } - // optional uint32 storeLimit = 8; - public static final int STORELIMIT_FIELD_NUMBER = 8; + // optional uint32 store_limit = 8; + public static final int STORE_LIMIT_FIELD_NUMBER = 8; private int storeLimit_; public boolean hasStoreLimit() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -691,8 +691,8 @@ public final class ClientProtos { return storeLimit_; } - // optional uint32 storeOffset = 9; - public static final int STOREOFFSET_FIELD_NUMBER = 9; + // optional uint32 store_offset = 9; + public static final int STORE_OFFSET_FIELD_NUMBER = 9; private int storeOffset_; public boolean hasStoreOffset() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -906,23 +906,23 @@ public final class ClientProtos { hash = (53 * hash) + getFilter().hashCode(); } if (hasTimeRange()) { - hash = (37 * hash) + TIMERANGE_FIELD_NUMBER; + hash = (37 * hash) + TIME_RANGE_FIELD_NUMBER; hash = (53 * hash) + getTimeRange().hashCode(); } if (hasMaxVersions()) { - hash = (37 * hash) + MAXVERSIONS_FIELD_NUMBER; + hash = (37 * hash) + MAX_VERSIONS_FIELD_NUMBER; hash = (53 * hash) + getMaxVersions(); } if (hasCacheBlocks()) { - hash = (37 * hash) + CACHEBLOCKS_FIELD_NUMBER; + hash = (37 * hash) + CACHE_BLOCKS_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getCacheBlocks()); } if (hasStoreLimit()) { - hash = (37 * hash) + STORELIMIT_FIELD_NUMBER; + hash = (37 * hash) + STORE_LIMIT_FIELD_NUMBER; hash = (53 * hash) + getStoreLimit(); } if (hasStoreOffset()) { - hash = (37 * hash) + STOREOFFSET_FIELD_NUMBER; + hash = (37 * hash) + STORE_OFFSET_FIELD_NUMBER; hash = (53 * hash) + getStoreOffset(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1860,7 +1860,7 @@ public final class ClientProtos { return filterBuilder_; } - // optional .TimeRange timeRange = 5; + // optional .TimeRange time_range = 5; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange timeRange_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder> timeRangeBuilder_; @@ -1950,7 +1950,7 @@ public final class ClientProtos { return timeRangeBuilder_; } - // optional uint32 maxVersions = 6 [default = 1]; + // optional uint32 max_versions = 6 [default = 1]; private int maxVersions_ = 1; public boolean hasMaxVersions() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -1971,7 +1971,7 @@ public final class ClientProtos { return this; } - // optional bool cacheBlocks = 7 [default = true]; + // optional bool cache_blocks = 7 [default = true]; private boolean cacheBlocks_ = true; public boolean hasCacheBlocks() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -1992,7 +1992,7 @@ public final class ClientProtos { return this; } - // optional uint32 storeLimit = 8; + // optional uint32 store_limit = 8; private int storeLimit_ ; public boolean hasStoreLimit() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -2013,7 +2013,7 @@ public final class ClientProtos { return this; } - // optional uint32 storeOffset = 9; + // optional uint32 store_offset = 9; private int storeOffset_ ; public boolean hasStoreOffset() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -2058,7 +2058,7 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellOrBuilder getCellOrBuilder( int index); - // optional int32 associatedCellCount = 2; + // optional int32 associated_cell_count = 2; boolean hasAssociatedCellCount(); int getAssociatedCellCount(); } @@ -2112,8 +2112,8 @@ public final class ClientProtos { return cell_.get(index); } - // optional int32 associatedCellCount = 2; - public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 2; + // optional int32 associated_cell_count = 2; + public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 2; private int associatedCellCount_; public boolean hasAssociatedCellCount() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2205,7 +2205,7 @@ public final class ClientProtos { hash = (53 * hash) + getCellList().hashCode(); } if (hasAssociatedCellCount()) { - hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER; + hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER; hash = (53 * hash) + getAssociatedCellCount(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2663,7 +2663,7 @@ public final class ClientProtos { return cellBuilder_; } - // optional int32 associatedCellCount = 2; + // optional int32 associated_cell_count = 2; private int associatedCellCount_ ; public boolean hasAssociatedCellCount() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2708,11 +2708,11 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get getGet(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder getGetOrBuilder(); - // optional bool closestRowBefore = 3; + // optional bool closest_row_before = 3; boolean hasClosestRowBefore(); boolean getClosestRowBefore(); - // optional bool existenceOnly = 4; + // optional bool existence_only = 4; boolean hasExistenceOnly(); boolean getExistenceOnly(); } @@ -2771,8 +2771,8 @@ public final class ClientProtos { return get_; } - // optional bool closestRowBefore = 3; - public static final int CLOSESTROWBEFORE_FIELD_NUMBER = 3; + // optional bool closest_row_before = 3; + public static final int CLOSEST_ROW_BEFORE_FIELD_NUMBER = 3; private boolean closestRowBefore_; public boolean hasClosestRowBefore() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2781,8 +2781,8 @@ public final class ClientProtos { return closestRowBefore_; } - // optional bool existenceOnly = 4; - public static final int EXISTENCEONLY_FIELD_NUMBER = 4; + // optional bool existence_only = 4; + public static final int EXISTENCE_ONLY_FIELD_NUMBER = 4; private boolean existenceOnly_; public boolean hasExistenceOnly() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2923,11 +2923,11 @@ public final class ClientProtos { hash = (53 * hash) + getGet().hashCode(); } if (hasClosestRowBefore()) { - hash = (37 * hash) + CLOSESTROWBEFORE_FIELD_NUMBER; + hash = (37 * hash) + CLOSEST_ROW_BEFORE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getClosestRowBefore()); } if (hasExistenceOnly()) { - hash = (37 * hash) + EXISTENCEONLY_FIELD_NUMBER; + hash = (37 * hash) + EXISTENCE_ONLY_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getExistenceOnly()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3415,7 +3415,7 @@ public final class ClientProtos { return getBuilder_; } - // optional bool closestRowBefore = 3; + // optional bool closest_row_before = 3; private boolean closestRowBefore_ ; public boolean hasClosestRowBefore() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -3436,7 +3436,7 @@ public final class ClientProtos { return this; } - // optional bool existenceOnly = 4; + // optional bool existence_only = 4; private boolean existenceOnly_ ; public boolean hasExistenceOnly() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -3486,11 +3486,11 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder getGetOrBuilder( int index); - // optional bool closestRowBefore = 3; + // optional bool closest_row_before = 3; boolean hasClosestRowBefore(); boolean getClosestRowBefore(); - // optional bool existenceOnly = 4; + // optional bool existence_only = 4; boolean hasExistenceOnly(); boolean getExistenceOnly(); } @@ -3557,8 +3557,8 @@ public final class ClientProtos { return get_.get(index); } - // optional bool closestRowBefore = 3; - public static final int CLOSESTROWBEFORE_FIELD_NUMBER = 3; + // optional bool closest_row_before = 3; + public static final int CLOSEST_ROW_BEFORE_FIELD_NUMBER = 3; private boolean closestRowBefore_; public boolean hasClosestRowBefore() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3567,8 +3567,8 @@ public final class ClientProtos { return closestRowBefore_; } - // optional bool existenceOnly = 4; - public static final int EXISTENCEONLY_FIELD_NUMBER = 4; + // optional bool existence_only = 4; + public static final int EXISTENCE_ONLY_FIELD_NUMBER = 4; private boolean existenceOnly_; public boolean hasExistenceOnly() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -3704,11 +3704,11 @@ public final class ClientProtos { hash = (53 * hash) + getGetList().hashCode(); } if (hasClosestRowBefore()) { - hash = (37 * hash) + CLOSESTROWBEFORE_FIELD_NUMBER; + hash = (37 * hash) + CLOSEST_ROW_BEFORE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getClosestRowBefore()); } if (hasExistenceOnly()) { - hash = (37 * hash) + EXISTENCEONLY_FIELD_NUMBER; + hash = (37 * hash) + EXISTENCE_ONLY_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getExistenceOnly()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -4311,7 +4311,7 @@ public final class ClientProtos { return getBuilder_; } - // optional bool closestRowBefore = 3; + // optional bool closest_row_before = 3; private boolean closestRowBefore_ ; public boolean hasClosestRowBefore() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -4332,7 +4332,7 @@ public final class ClientProtos { return this; } - // optional bool existenceOnly = 4; + // optional bool existence_only = 4; private boolean existenceOnly_ ; public boolean hasExistenceOnly() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -5594,7 +5594,7 @@ public final class ClientProtos { boolean hasQualifier(); com.google.protobuf.ByteString getQualifier(); - // required .CompareType compareType = 4; + // required .CompareType compare_type = 4; boolean hasCompareType(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareType(); @@ -5662,8 +5662,8 @@ public final class ClientProtos { return qualifier_; } - // required .CompareType compareType = 4; - public static final int COMPARETYPE_FIELD_NUMBER = 4; + // required .CompareType compare_type = 4; + public static final int COMPARE_TYPE_FIELD_NUMBER = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareType_; public boolean hasCompareType() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -5842,7 +5842,7 @@ public final class ClientProtos { hash = (53 * hash) + getQualifier().hashCode(); } if (hasCompareType()) { - hash = (37 * hash) + COMPARETYPE_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_TYPE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getCompareType()); } if (hasComparator()) { @@ -6241,7 +6241,7 @@ public final class ClientProtos { return this; } - // required .CompareType compareType = 4; + // required .CompareType compare_type = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareType_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS; public boolean hasCompareType() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -6373,11 +6373,11 @@ public final class ClientProtos { boolean hasRow(); com.google.protobuf.ByteString getRow(); - // optional .MutationProto.MutationType mutateType = 2; + // optional .MutationProto.MutationType mutate_type = 2; boolean hasMutateType(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType getMutateType(); - // repeated .MutationProto.ColumnValue columnValue = 3; + // repeated .MutationProto.ColumnValue column_value = 3; java.util.List getColumnValueList(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.ColumnValue getColumnValue(int index); @@ -6405,12 +6405,12 @@ public final class ClientProtos { boolean hasDurability(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Durability getDurability(); - // optional .TimeRange timeRange = 7; + // optional .TimeRange time_range = 7; boolean hasTimeRange(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange getTimeRange(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder getTimeRangeOrBuilder(); - // optional int32 associatedCellCount = 8; + // optional int32 associated_cell_count = 8; boolean hasAssociatedCellCount(); int getAssociatedCellCount(); } @@ -6677,7 +6677,7 @@ public final class ClientProtos { boolean hasFamily(); com.google.protobuf.ByteString getFamily(); - // repeated .MutationProto.ColumnValue.QualifierValue qualifierValue = 2; + // repeated .MutationProto.ColumnValue.QualifierValue qualifier_value = 2; java.util.List getQualifierValueList(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.ColumnValue.QualifierValue getQualifierValue(int index); @@ -6730,7 +6730,7 @@ public final class ClientProtos { boolean hasTimestamp(); long getTimestamp(); - // optional .MutationProto.DeleteType deleteType = 4; + // optional .MutationProto.DeleteType delete_type = 4; boolean hasDeleteType(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.DeleteType getDeleteType(); } @@ -6793,8 +6793,8 @@ public final class ClientProtos { return timestamp_; } - // optional .MutationProto.DeleteType deleteType = 4; - public static final int DELETETYPE_FIELD_NUMBER = 4; + // optional .MutationProto.DeleteType delete_type = 4; + public static final int DELETE_TYPE_FIELD_NUMBER = 4; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.DeleteType deleteType_; public boolean hasDeleteType() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -6923,7 +6923,7 @@ public final class ClientProtos { hash = (53 * hash) + hashLong(getTimestamp()); } if (hasDeleteType()) { - hash = (37 * hash) + DELETETYPE_FIELD_NUMBER; + hash = (37 * hash) + DELETE_TYPE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getDeleteType()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -7264,7 +7264,7 @@ public final class ClientProtos { return this; } - // optional .MutationProto.DeleteType deleteType = 4; + // optional .MutationProto.DeleteType delete_type = 4; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.DeleteType deleteType_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.DeleteType.DELETE_ONE_VERSION; public boolean hasDeleteType() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -7310,8 +7310,8 @@ public final class ClientProtos { return family_; } - // repeated .MutationProto.ColumnValue.QualifierValue qualifierValue = 2; - public static final int QUALIFIERVALUE_FIELD_NUMBER = 2; + // repeated .MutationProto.ColumnValue.QualifierValue qualifier_value = 2; + public static final int QUALIFIER_VALUE_FIELD_NUMBER = 2; private java.util.List qualifierValue_; public java.util.List getQualifierValueList() { return qualifierValue_; @@ -7418,7 +7418,7 @@ public final class ClientProtos { hash = (53 * hash) + getFamily().hashCode(); } if (getQualifierValueCount() > 0) { - hash = (37 * hash) + QUALIFIERVALUE_FIELD_NUMBER; + hash = (37 * hash) + QUALIFIER_VALUE_FIELD_NUMBER; hash = (53 * hash) + getQualifierValueList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -7718,7 +7718,7 @@ public final class ClientProtos { return this; } - // repeated .MutationProto.ColumnValue.QualifierValue qualifierValue = 2; + // repeated .MutationProto.ColumnValue.QualifierValue qualifier_value = 2; private java.util.List qualifierValue_ = java.util.Collections.emptyList(); private void ensureQualifierValueIsMutable() { @@ -7926,8 +7926,8 @@ public final class ClientProtos { return row_; } - // optional .MutationProto.MutationType mutateType = 2; - public static final int MUTATETYPE_FIELD_NUMBER = 2; + // optional .MutationProto.MutationType mutate_type = 2; + public static final int MUTATE_TYPE_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType mutateType_; public boolean hasMutateType() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -7936,8 +7936,8 @@ public final class ClientProtos { return mutateType_; } - // repeated .MutationProto.ColumnValue columnValue = 3; - public static final int COLUMNVALUE_FIELD_NUMBER = 3; + // repeated .MutationProto.ColumnValue column_value = 3; + public static final int COLUMN_VALUE_FIELD_NUMBER = 3; private java.util.List columnValue_; public java.util.List getColumnValueList() { return columnValue_; @@ -7998,8 +7998,8 @@ public final class ClientProtos { return durability_; } - // optional .TimeRange timeRange = 7; - public static final int TIMERANGE_FIELD_NUMBER = 7; + // optional .TimeRange time_range = 7; + public static final int TIME_RANGE_FIELD_NUMBER = 7; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange timeRange_; public boolean hasTimeRange() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -8011,8 +8011,8 @@ public final class ClientProtos { return timeRange_; } - // optional int32 associatedCellCount = 8; - public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 8; + // optional int32 associated_cell_count = 8; + public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 8; private int associatedCellCount_; public boolean hasAssociatedCellCount() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -8191,11 +8191,11 @@ public final class ClientProtos { hash = (53 * hash) + getRow().hashCode(); } if (hasMutateType()) { - hash = (37 * hash) + MUTATETYPE_FIELD_NUMBER; + hash = (37 * hash) + MUTATE_TYPE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getMutateType()); } if (getColumnValueCount() > 0) { - hash = (37 * hash) + COLUMNVALUE_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_VALUE_FIELD_NUMBER; hash = (53 * hash) + getColumnValueList().hashCode(); } if (hasTimestamp()) { @@ -8211,11 +8211,11 @@ public final class ClientProtos { hash = (53 * hash) + hashEnum(getDurability()); } if (hasTimeRange()) { - hash = (37 * hash) + TIMERANGE_FIELD_NUMBER; + hash = (37 * hash) + TIME_RANGE_FIELD_NUMBER; hash = (53 * hash) + getTimeRange().hashCode(); } if (hasAssociatedCellCount()) { - hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER; + hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER; hash = (53 * hash) + getAssociatedCellCount(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -8666,7 +8666,7 @@ public final class ClientProtos { return this; } - // optional .MutationProto.MutationType mutateType = 2; + // optional .MutationProto.MutationType mutate_type = 2; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType mutateType_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType.APPEND; public boolean hasMutateType() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -8690,7 +8690,7 @@ public final class ClientProtos { return this; } - // repeated .MutationProto.ColumnValue columnValue = 3; + // repeated .MutationProto.ColumnValue column_value = 3; private java.util.List columnValue_ = java.util.Collections.emptyList(); private void ensureColumnValueIsMutable() { @@ -9107,7 +9107,7 @@ public final class ClientProtos { return this; } - // optional .TimeRange timeRange = 7; + // optional .TimeRange time_range = 7; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange timeRange_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder> timeRangeBuilder_; @@ -9197,7 +9197,7 @@ public final class ClientProtos { return timeRangeBuilder_; } - // optional int32 associatedCellCount = 8; + // optional int32 associated_cell_count = 8; private int associatedCellCount_ ; public boolean hasAssociatedCellCount() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -10580,11 +10580,11 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getAttributeOrBuilder( int index); - // optional bytes startRow = 3; + // optional bytes start_row = 3; boolean hasStartRow(); com.google.protobuf.ByteString getStartRow(); - // optional bytes stopRow = 4; + // optional bytes stop_row = 4; boolean hasStopRow(); com.google.protobuf.ByteString getStopRow(); @@ -10593,40 +10593,40 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter getFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.FilterOrBuilder getFilterOrBuilder(); - // optional .TimeRange timeRange = 6; + // optional .TimeRange time_range = 6; boolean hasTimeRange(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange getTimeRange(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder getTimeRangeOrBuilder(); - // optional uint32 maxVersions = 7 [default = 1]; + // optional uint32 max_versions = 7 [default = 1]; boolean hasMaxVersions(); int getMaxVersions(); - // optional bool cacheBlocks = 8 [default = true]; + // optional bool cache_blocks = 8 [default = true]; boolean hasCacheBlocks(); boolean getCacheBlocks(); - // optional uint32 batchSize = 9; + // optional uint32 batch_size = 9; boolean hasBatchSize(); int getBatchSize(); - // optional uint64 maxResultSize = 10; + // optional uint64 max_result_size = 10; boolean hasMaxResultSize(); long getMaxResultSize(); - // optional uint32 storeLimit = 11; + // optional uint32 store_limit = 11; boolean hasStoreLimit(); int getStoreLimit(); - // optional uint32 storeOffset = 12; + // optional uint32 store_offset = 12; boolean hasStoreOffset(); int getStoreOffset(); - // optional bool loadColumnFamiliesOnDemand = 13; + // optional bool load_column_families_on_demand = 13; boolean hasLoadColumnFamiliesOnDemand(); boolean getLoadColumnFamiliesOnDemand(); - // optional uint32 cachingCount = 14; + // optional uint32 caching_count = 14; boolean hasCachingCount(); int getCachingCount(); @@ -10705,8 +10705,8 @@ public final class ClientProtos { return attribute_.get(index); } - // optional bytes startRow = 3; - public static final int STARTROW_FIELD_NUMBER = 3; + // optional bytes start_row = 3; + public static final int START_ROW_FIELD_NUMBER = 3; private com.google.protobuf.ByteString startRow_; public boolean hasStartRow() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -10715,8 +10715,8 @@ public final class ClientProtos { return startRow_; } - // optional bytes stopRow = 4; - public static final int STOPROW_FIELD_NUMBER = 4; + // optional bytes stop_row = 4; + public static final int STOP_ROW_FIELD_NUMBER = 4; private com.google.protobuf.ByteString stopRow_; public boolean hasStopRow() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -10738,8 +10738,8 @@ public final class ClientProtos { return filter_; } - // optional .TimeRange timeRange = 6; - public static final int TIMERANGE_FIELD_NUMBER = 6; + // optional .TimeRange time_range = 6; + public static final int TIME_RANGE_FIELD_NUMBER = 6; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange timeRange_; public boolean hasTimeRange() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -10751,8 +10751,8 @@ public final class ClientProtos { return timeRange_; } - // optional uint32 maxVersions = 7 [default = 1]; - public static final int MAXVERSIONS_FIELD_NUMBER = 7; + // optional uint32 max_versions = 7 [default = 1]; + public static final int MAX_VERSIONS_FIELD_NUMBER = 7; private int maxVersions_; public boolean hasMaxVersions() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -10761,8 +10761,8 @@ public final class ClientProtos { return maxVersions_; } - // optional bool cacheBlocks = 8 [default = true]; - public static final int CACHEBLOCKS_FIELD_NUMBER = 8; + // optional bool cache_blocks = 8 [default = true]; + public static final int CACHE_BLOCKS_FIELD_NUMBER = 8; private boolean cacheBlocks_; public boolean hasCacheBlocks() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -10771,8 +10771,8 @@ public final class ClientProtos { return cacheBlocks_; } - // optional uint32 batchSize = 9; - public static final int BATCHSIZE_FIELD_NUMBER = 9; + // optional uint32 batch_size = 9; + public static final int BATCH_SIZE_FIELD_NUMBER = 9; private int batchSize_; public boolean hasBatchSize() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -10781,8 +10781,8 @@ public final class ClientProtos { return batchSize_; } - // optional uint64 maxResultSize = 10; - public static final int MAXRESULTSIZE_FIELD_NUMBER = 10; + // optional uint64 max_result_size = 10; + public static final int MAX_RESULT_SIZE_FIELD_NUMBER = 10; private long maxResultSize_; public boolean hasMaxResultSize() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -10791,8 +10791,8 @@ public final class ClientProtos { return maxResultSize_; } - // optional uint32 storeLimit = 11; - public static final int STORELIMIT_FIELD_NUMBER = 11; + // optional uint32 store_limit = 11; + public static final int STORE_LIMIT_FIELD_NUMBER = 11; private int storeLimit_; public boolean hasStoreLimit() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -10801,8 +10801,8 @@ public final class ClientProtos { return storeLimit_; } - // optional uint32 storeOffset = 12; - public static final int STOREOFFSET_FIELD_NUMBER = 12; + // optional uint32 store_offset = 12; + public static final int STORE_OFFSET_FIELD_NUMBER = 12; private int storeOffset_; public boolean hasStoreOffset() { return ((bitField0_ & 0x00000200) == 0x00000200); @@ -10811,8 +10811,8 @@ public final class ClientProtos { return storeOffset_; } - // optional bool loadColumnFamiliesOnDemand = 13; - public static final int LOADCOLUMNFAMILIESONDEMAND_FIELD_NUMBER = 13; + // optional bool load_column_families_on_demand = 13; + public static final int LOAD_COLUMN_FAMILIES_ON_DEMAND_FIELD_NUMBER = 13; private boolean loadColumnFamiliesOnDemand_; public boolean hasLoadColumnFamiliesOnDemand() { return ((bitField0_ & 0x00000400) == 0x00000400); @@ -10821,8 +10821,8 @@ public final class ClientProtos { return loadColumnFamiliesOnDemand_; } - // optional uint32 cachingCount = 14; - public static final int CACHINGCOUNT_FIELD_NUMBER = 14; + // optional uint32 caching_count = 14; + public static final int CACHING_COUNT_FIELD_NUMBER = 14; private int cachingCount_; public boolean hasCachingCount() { return ((bitField0_ & 0x00000800) == 0x00000800); @@ -11112,11 +11112,11 @@ public final class ClientProtos { hash = (53 * hash) + getAttributeList().hashCode(); } if (hasStartRow()) { - hash = (37 * hash) + STARTROW_FIELD_NUMBER; + hash = (37 * hash) + START_ROW_FIELD_NUMBER; hash = (53 * hash) + getStartRow().hashCode(); } if (hasStopRow()) { - hash = (37 * hash) + STOPROW_FIELD_NUMBER; + hash = (37 * hash) + STOP_ROW_FIELD_NUMBER; hash = (53 * hash) + getStopRow().hashCode(); } if (hasFilter()) { @@ -11124,39 +11124,39 @@ public final class ClientProtos { hash = (53 * hash) + getFilter().hashCode(); } if (hasTimeRange()) { - hash = (37 * hash) + TIMERANGE_FIELD_NUMBER; + hash = (37 * hash) + TIME_RANGE_FIELD_NUMBER; hash = (53 * hash) + getTimeRange().hashCode(); } if (hasMaxVersions()) { - hash = (37 * hash) + MAXVERSIONS_FIELD_NUMBER; + hash = (37 * hash) + MAX_VERSIONS_FIELD_NUMBER; hash = (53 * hash) + getMaxVersions(); } if (hasCacheBlocks()) { - hash = (37 * hash) + CACHEBLOCKS_FIELD_NUMBER; + hash = (37 * hash) + CACHE_BLOCKS_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getCacheBlocks()); } if (hasBatchSize()) { - hash = (37 * hash) + BATCHSIZE_FIELD_NUMBER; + hash = (37 * hash) + BATCH_SIZE_FIELD_NUMBER; hash = (53 * hash) + getBatchSize(); } if (hasMaxResultSize()) { - hash = (37 * hash) + MAXRESULTSIZE_FIELD_NUMBER; + hash = (37 * hash) + MAX_RESULT_SIZE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getMaxResultSize()); } if (hasStoreLimit()) { - hash = (37 * hash) + STORELIMIT_FIELD_NUMBER; + hash = (37 * hash) + STORE_LIMIT_FIELD_NUMBER; hash = (53 * hash) + getStoreLimit(); } if (hasStoreOffset()) { - hash = (37 * hash) + STOREOFFSET_FIELD_NUMBER; + hash = (37 * hash) + STORE_OFFSET_FIELD_NUMBER; hash = (53 * hash) + getStoreOffset(); } if (hasLoadColumnFamiliesOnDemand()) { - hash = (37 * hash) + LOADCOLUMNFAMILIESONDEMAND_FIELD_NUMBER; + hash = (37 * hash) + LOAD_COLUMN_FAMILIES_ON_DEMAND_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getLoadColumnFamiliesOnDemand()); } if (hasCachingCount()) { - hash = (37 * hash) + CACHINGCOUNT_FIELD_NUMBER; + hash = (37 * hash) + CACHING_COUNT_FIELD_NUMBER; hash = (53 * hash) + getCachingCount(); } if (hasPrefetching()) { @@ -12064,7 +12064,7 @@ public final class ClientProtos { return attributeBuilder_; } - // optional bytes startRow = 3; + // optional bytes start_row = 3; private com.google.protobuf.ByteString startRow_ = com.google.protobuf.ByteString.EMPTY; public boolean hasStartRow() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -12088,7 +12088,7 @@ public final class ClientProtos { return this; } - // optional bytes stopRow = 4; + // optional bytes stop_row = 4; private com.google.protobuf.ByteString stopRow_ = com.google.protobuf.ByteString.EMPTY; public boolean hasStopRow() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -12202,7 +12202,7 @@ public final class ClientProtos { return filterBuilder_; } - // optional .TimeRange timeRange = 6; + // optional .TimeRange time_range = 6; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange timeRange_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRange.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeRangeOrBuilder> timeRangeBuilder_; @@ -12292,7 +12292,7 @@ public final class ClientProtos { return timeRangeBuilder_; } - // optional uint32 maxVersions = 7 [default = 1]; + // optional uint32 max_versions = 7 [default = 1]; private int maxVersions_ = 1; public boolean hasMaxVersions() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -12313,7 +12313,7 @@ public final class ClientProtos { return this; } - // optional bool cacheBlocks = 8 [default = true]; + // optional bool cache_blocks = 8 [default = true]; private boolean cacheBlocks_ = true; public boolean hasCacheBlocks() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -12334,7 +12334,7 @@ public final class ClientProtos { return this; } - // optional uint32 batchSize = 9; + // optional uint32 batch_size = 9; private int batchSize_ ; public boolean hasBatchSize() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -12355,7 +12355,7 @@ public final class ClientProtos { return this; } - // optional uint64 maxResultSize = 10; + // optional uint64 max_result_size = 10; private long maxResultSize_ ; public boolean hasMaxResultSize() { return ((bitField0_ & 0x00000200) == 0x00000200); @@ -12376,7 +12376,7 @@ public final class ClientProtos { return this; } - // optional uint32 storeLimit = 11; + // optional uint32 store_limit = 11; private int storeLimit_ ; public boolean hasStoreLimit() { return ((bitField0_ & 0x00000400) == 0x00000400); @@ -12397,7 +12397,7 @@ public final class ClientProtos { return this; } - // optional uint32 storeOffset = 12; + // optional uint32 store_offset = 12; private int storeOffset_ ; public boolean hasStoreOffset() { return ((bitField0_ & 0x00000800) == 0x00000800); @@ -12418,7 +12418,7 @@ public final class ClientProtos { return this; } - // optional bool loadColumnFamiliesOnDemand = 13; + // optional bool load_column_families_on_demand = 13; private boolean loadColumnFamiliesOnDemand_ ; public boolean hasLoadColumnFamiliesOnDemand() { return ((bitField0_ & 0x00001000) == 0x00001000); @@ -12439,7 +12439,7 @@ public final class ClientProtos { return this; } - // optional uint32 cachingCount = 14; + // optional uint32 caching_count = 14; private int cachingCount_ ; public boolean hasCachingCount() { return ((bitField0_ & 0x00002000) == 0x00002000); @@ -12505,19 +12505,19 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder(); - // optional uint64 scannerId = 3; + // optional uint64 scanner_id = 3; boolean hasScannerId(); long getScannerId(); - // optional uint32 numberOfRows = 4; + // optional uint32 number_of_rows = 4; boolean hasNumberOfRows(); int getNumberOfRows(); - // optional bool closeScanner = 5; + // optional bool close_scanner = 5; boolean hasCloseScanner(); boolean getCloseScanner(); - // optional uint64 nextCallSeq = 6; + // optional uint64 next_call_seq = 6; boolean hasNextCallSeq(); long getNextCallSeq(); } @@ -12576,8 +12576,8 @@ public final class ClientProtos { return scan_; } - // optional uint64 scannerId = 3; - public static final int SCANNERID_FIELD_NUMBER = 3; + // optional uint64 scanner_id = 3; + public static final int SCANNER_ID_FIELD_NUMBER = 3; private long scannerId_; public boolean hasScannerId() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -12586,8 +12586,8 @@ public final class ClientProtos { return scannerId_; } - // optional uint32 numberOfRows = 4; - public static final int NUMBEROFROWS_FIELD_NUMBER = 4; + // optional uint32 number_of_rows = 4; + public static final int NUMBER_OF_ROWS_FIELD_NUMBER = 4; private int numberOfRows_; public boolean hasNumberOfRows() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -12596,8 +12596,8 @@ public final class ClientProtos { return numberOfRows_; } - // optional bool closeScanner = 5; - public static final int CLOSESCANNER_FIELD_NUMBER = 5; + // optional bool close_scanner = 5; + public static final int CLOSE_SCANNER_FIELD_NUMBER = 5; private boolean closeScanner_; public boolean hasCloseScanner() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -12606,8 +12606,8 @@ public final class ClientProtos { return closeScanner_; } - // optional uint64 nextCallSeq = 6; - public static final int NEXTCALLSEQ_FIELD_NUMBER = 6; + // optional uint64 next_call_seq = 6; + public static final int NEXT_CALL_SEQ_FIELD_NUMBER = 6; private long nextCallSeq_; public boolean hasNextCallSeq() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -12770,19 +12770,19 @@ public final class ClientProtos { hash = (53 * hash) + getScan().hashCode(); } if (hasScannerId()) { - hash = (37 * hash) + SCANNERID_FIELD_NUMBER; + hash = (37 * hash) + SCANNER_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getScannerId()); } if (hasNumberOfRows()) { - hash = (37 * hash) + NUMBEROFROWS_FIELD_NUMBER; + hash = (37 * hash) + NUMBER_OF_ROWS_FIELD_NUMBER; hash = (53 * hash) + getNumberOfRows(); } if (hasCloseScanner()) { - hash = (37 * hash) + CLOSESCANNER_FIELD_NUMBER; + hash = (37 * hash) + CLOSE_SCANNER_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getCloseScanner()); } if (hasNextCallSeq()) { - hash = (37 * hash) + NEXTCALLSEQ_FIELD_NUMBER; + hash = (37 * hash) + NEXT_CALL_SEQ_FIELD_NUMBER; hash = (53 * hash) + hashLong(getNextCallSeq()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -13294,7 +13294,7 @@ public final class ClientProtos { return scanBuilder_; } - // optional uint64 scannerId = 3; + // optional uint64 scanner_id = 3; private long scannerId_ ; public boolean hasScannerId() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -13315,7 +13315,7 @@ public final class ClientProtos { return this; } - // optional uint32 numberOfRows = 4; + // optional uint32 number_of_rows = 4; private int numberOfRows_ ; public boolean hasNumberOfRows() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -13336,7 +13336,7 @@ public final class ClientProtos { return this; } - // optional bool closeScanner = 5; + // optional bool close_scanner = 5; private boolean closeScanner_ ; public boolean hasCloseScanner() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -13357,7 +13357,7 @@ public final class ClientProtos { return this; } - // optional uint64 nextCallSeq = 6; + // optional uint64 next_call_seq = 6; private long nextCallSeq_ ; public boolean hasNextCallSeq() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -13392,16 +13392,16 @@ public final class ClientProtos { public interface ScanResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional .ResultCellMeta resultCellMeta = 1; + // optional .ResultCellMeta result_cell_meta = 1; boolean hasResultCellMeta(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta getResultCellMeta(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMetaOrBuilder getResultCellMetaOrBuilder(); - // optional uint64 scannerId = 2; + // optional uint64 scanner_id = 2; boolean hasScannerId(); long getScannerId(); - // optional bool moreResults = 3; + // optional bool more_results = 3; boolean hasMoreResults(); boolean getMoreResults(); @@ -13438,8 +13438,8 @@ public final class ClientProtos { } private int bitField0_; - // optional .ResultCellMeta resultCellMeta = 1; - public static final int RESULTCELLMETA_FIELD_NUMBER = 1; + // optional .ResultCellMeta result_cell_meta = 1; + public static final int RESULT_CELL_META_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta resultCellMeta_; public boolean hasResultCellMeta() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -13451,8 +13451,8 @@ public final class ClientProtos { return resultCellMeta_; } - // optional uint64 scannerId = 2; - public static final int SCANNERID_FIELD_NUMBER = 2; + // optional uint64 scanner_id = 2; + public static final int SCANNER_ID_FIELD_NUMBER = 2; private long scannerId_; public boolean hasScannerId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -13461,8 +13461,8 @@ public final class ClientProtos { return scannerId_; } - // optional bool moreResults = 3; - public static final int MORERESULTS_FIELD_NUMBER = 3; + // optional bool more_results = 3; + public static final int MORE_RESULTS_FIELD_NUMBER = 3; private boolean moreResults_; public boolean hasMoreResults() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -13589,15 +13589,15 @@ public final class ClientProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasResultCellMeta()) { - hash = (37 * hash) + RESULTCELLMETA_FIELD_NUMBER; + hash = (37 * hash) + RESULT_CELL_META_FIELD_NUMBER; hash = (53 * hash) + getResultCellMeta().hashCode(); } if (hasScannerId()) { - hash = (37 * hash) + SCANNERID_FIELD_NUMBER; + hash = (37 * hash) + SCANNER_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getScannerId()); } if (hasMoreResults()) { - hash = (37 * hash) + MORERESULTS_FIELD_NUMBER; + hash = (37 * hash) + MORE_RESULTS_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getMoreResults()); } if (hasTtl()) { @@ -13880,7 +13880,7 @@ public final class ClientProtos { private int bitField0_; - // optional .ResultCellMeta resultCellMeta = 1; + // optional .ResultCellMeta result_cell_meta = 1; private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta resultCellMeta_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultCellMetaOrBuilder> resultCellMetaBuilder_; @@ -13970,7 +13970,7 @@ public final class ClientProtos { return resultCellMetaBuilder_; } - // optional uint64 scannerId = 2; + // optional uint64 scanner_id = 2; private long scannerId_ ; public boolean hasScannerId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -13991,7 +13991,7 @@ public final class ClientProtos { return this; } - // optional bool moreResults = 3; + // optional bool more_results = 3; private boolean moreResults_ ; public boolean hasMoreResults() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -14047,7 +14047,7 @@ public final class ClientProtos { public interface ResultCellMetaOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated uint32 cellsLength = 1; + // repeated uint32 cells_length = 1; java.util.List getCellsLengthList(); int getCellsLengthCount(); int getCellsLength(int index); @@ -14080,8 +14080,8 @@ public final class ClientProtos { return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ResultCellMeta_fieldAccessorTable; } - // repeated uint32 cellsLength = 1; - public static final int CELLSLENGTH_FIELD_NUMBER = 1; + // repeated uint32 cells_length = 1; + public static final int CELLS_LENGTH_FIELD_NUMBER = 1; private java.util.List cellsLength_; public java.util.List getCellsLengthList() { @@ -14165,7 +14165,7 @@ public final class ClientProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getCellsLengthCount() > 0) { - hash = (37 * hash) + CELLSLENGTH_FIELD_NUMBER; + hash = (37 * hash) + CELLS_LENGTH_FIELD_NUMBER; hash = (53 * hash) + getCellsLengthList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -14404,7 +14404,7 @@ public final class ClientProtos { private int bitField0_; - // repeated uint32 cellsLength = 1; + // repeated uint32 cells_length = 1; private java.util.List cellsLength_ = java.util.Collections.emptyList();; private void ensureCellsLengthIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { @@ -14468,7 +14468,7 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); - // repeated .BulkLoadHFileRequest.FamilyPath familyPath = 2; + // repeated .BulkLoadHFileRequest.FamilyPath family_path = 2; java.util.List getFamilyPathList(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index); @@ -14478,7 +14478,7 @@ public final class ClientProtos { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder( int index); - // optional bool assignSeqNum = 3; + // optional bool assign_seq_num = 3; boolean hasAssignSeqNum(); boolean getAssignSeqNum(); } @@ -15017,8 +15017,8 @@ public final class ClientProtos { return region_; } - // repeated .BulkLoadHFileRequest.FamilyPath familyPath = 2; - public static final int FAMILYPATH_FIELD_NUMBER = 2; + // repeated .BulkLoadHFileRequest.FamilyPath family_path = 2; + public static final int FAMILY_PATH_FIELD_NUMBER = 2; private java.util.List familyPath_; public java.util.List getFamilyPathList() { return familyPath_; @@ -15038,8 +15038,8 @@ public final class ClientProtos { return familyPath_.get(index); } - // optional bool assignSeqNum = 3; - public static final int ASSIGNSEQNUM_FIELD_NUMBER = 3; + // optional bool assign_seq_num = 3; + public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 3; private boolean assignSeqNum_; public boolean hasAssignSeqNum() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -15158,11 +15158,11 @@ public final class ClientProtos { hash = (53 * hash) + getRegion().hashCode(); } if (getFamilyPathCount() > 0) { - hash = (37 * hash) + FAMILYPATH_FIELD_NUMBER; + hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER; hash = (53 * hash) + getFamilyPathList().hashCode(); } if (hasAssignSeqNum()) { - hash = (37 * hash) + ASSIGNSEQNUM_FIELD_NUMBER; + hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getAssignSeqNum()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -15565,7 +15565,7 @@ public final class ClientProtos { return regionBuilder_; } - // repeated .BulkLoadHFileRequest.FamilyPath familyPath = 2; + // repeated .BulkLoadHFileRequest.FamilyPath family_path = 2; private java.util.List familyPath_ = java.util.Collections.emptyList(); private void ensureFamilyPathIsMutable() { @@ -15751,7 +15751,7 @@ public final class ClientProtos { return familyPathBuilder_; } - // optional bool assignSeqNum = 3; + // optional bool assign_seq_num = 3; private boolean assignSeqNum_ ; public boolean hasAssignSeqNum() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -16169,11 +16169,11 @@ public final class ClientProtos { boolean hasRow(); com.google.protobuf.ByteString getRow(); - // required string serviceName = 2; + // required string service_name = 2; boolean hasServiceName(); String getServiceName(); - // required string methodName = 3; + // required string method_name = 3; boolean hasMethodName(); String getMethodName(); @@ -16220,8 +16220,8 @@ public final class ClientProtos { return row_; } - // required string serviceName = 2; - public static final int SERVICENAME_FIELD_NUMBER = 2; + // required string service_name = 2; + public static final int SERVICE_NAME_FIELD_NUMBER = 2; private java.lang.Object serviceName_; public boolean hasServiceName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -16252,8 +16252,8 @@ public final class ClientProtos { } } - // required string methodName = 3; - public static final int METHODNAME_FIELD_NUMBER = 3; + // required string method_name = 3; + public static final int METHOD_NAME_FIELD_NUMBER = 3; private java.lang.Object methodName_; public boolean hasMethodName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -16422,11 +16422,11 @@ public final class ClientProtos { hash = (53 * hash) + getRow().hashCode(); } if (hasServiceName()) { - hash = (37 * hash) + SERVICENAME_FIELD_NUMBER; + hash = (37 * hash) + SERVICE_NAME_FIELD_NUMBER; hash = (53 * hash) + getServiceName().hashCode(); } if (hasMethodName()) { - hash = (37 * hash) + METHODNAME_FIELD_NUMBER; + hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER; hash = (53 * hash) + getMethodName().hashCode(); } if (hasRequest()) { @@ -16736,7 +16736,7 @@ public final class ClientProtos { return this; } - // required string serviceName = 2; + // required string service_name = 2; private java.lang.Object serviceName_ = ""; public boolean hasServiceName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -16772,7 +16772,7 @@ public final class ClientProtos { onChanged(); } - // required string methodName = 3; + // required string method_name = 3; private java.lang.Object methodName_ = ""; public boolean hasMethodName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -21559,98 +21559,100 @@ public final class ClientProtos { java.lang.String[] descriptorData = { "\n\014Client.proto\032\013hbase.proto\032\014Filter.prot" + "o\032\nCell.proto\032\020Comparator.proto\"+\n\006Colum" + - "n\022\016\n\006family\030\001 \002(\014\022\021\n\tqualifier\030\002 \003(\014\"\342\001\n" + + "n\022\016\n\006family\030\001 \002(\014\022\021\n\tqualifier\030\002 \003(\014\"\347\001\n" + "\003Get\022\013\n\003row\030\001 \002(\014\022\027\n\006column\030\002 \003(\0132\007.Colu" + "mn\022!\n\tattribute\030\003 \003(\0132\016.NameBytesPair\022\027\n" + - "\006filter\030\004 \001(\0132\007.Filter\022\035\n\ttimeRange\030\005 \001(" + - "\0132\n.TimeRange\022\026\n\013maxVersions\030\006 \001(\r:\0011\022\031\n" + - "\013cacheBlocks\030\007 \001(\010:\004true\022\022\n\nstoreLimit\030\010" + - " \001(\r\022\023\n\013storeOffset\030\t \001(\r\":\n\006Result\022\023\n\004c" + - "ell\030\001 \003(\0132\005.Cell\022\033\n\023associatedCellCount\030", - "\002 \001(\005\"r\n\nGetRequest\022 \n\006region\030\001 \002(\0132\020.Re" + - "gionSpecifier\022\021\n\003get\030\002 \002(\0132\004.Get\022\030\n\020clos" + - "estRowBefore\030\003 \001(\010\022\025\n\rexistenceOnly\030\004 \001(" + - "\010\"w\n\017MultiGetRequest\022 \n\006region\030\001 \002(\0132\020.R" + - "egionSpecifier\022\021\n\003get\030\002 \003(\0132\004.Get\022\030\n\020clo" + - "sestRowBefore\030\003 \001(\010\022\025\n\rexistenceOnly\030\004 \001" + - "(\010\"6\n\013GetResponse\022\027\n\006result\030\001 \001(\0132\007.Resu" + - "lt\022\016\n\006exists\030\002 \001(\010\";\n\020MultiGetResponse\022\027" + - "\n\006result\030\001 \003(\0132\007.Result\022\016\n\006exists\030\002 \003(\010\"" + - "\177\n\tCondition\022\013\n\003row\030\001 \002(\014\022\016\n\006family\030\002 \002(", - "\014\022\021\n\tqualifier\030\003 \002(\014\022!\n\013compareType\030\004 \002(" + - "\0162\014.CompareType\022\037\n\ncomparator\030\005 \002(\0132\013.Co" + - "mparator\"\220\006\n\rMutationProto\022\013\n\003row\030\001 \001(\014\022" + - "/\n\nmutateType\030\002 \001(\0162\033.MutationProto.Muta" + - "tionType\022/\n\013columnValue\030\003 \003(\0132\032.Mutation" + - "Proto.ColumnValue\022\021\n\ttimestamp\030\004 \001(\004\022!\n\t" + - "attribute\030\005 \003(\0132\016.NameBytesPair\022:\n\ndurab" + - "ility\030\006 \001(\0162\031.MutationProto.Durability:\013" + - "USE_DEFAULT\022\035\n\ttimeRange\030\007 \001(\0132\n.TimeRan" + - "ge\022\033\n\023associatedCellCount\030\010 \001(\005\032\326\001\n\013Colu", - "mnValue\022\016\n\006family\030\001 \002(\014\022A\n\016qualifierValu" + - "e\030\002 \003(\0132).MutationProto.ColumnValue.Qual" + - "ifierValue\032t\n\016QualifierValue\022\021\n\tqualifie" + - "r\030\001 \001(\014\022\r\n\005value\030\002 \001(\014\022\021\n\ttimestamp\030\003 \001(" + - "\004\022-\n\ndeleteType\030\004 \001(\0162\031.MutationProto.De" + - "leteType\"W\n\nDurability\022\017\n\013USE_DEFAULT\020\000\022" + - "\014\n\010SKIP_WAL\020\001\022\r\n\tASYNC_WAL\020\002\022\014\n\010SYNC_WAL" + - "\020\003\022\r\n\tFSYNC_WAL\020\004\">\n\014MutationType\022\n\n\006APP" + - "END\020\000\022\r\n\tINCREMENT\020\001\022\007\n\003PUT\020\002\022\n\n\006DELETE\020" + - "\003\"p\n\nDeleteType\022\026\n\022DELETE_ONE_VERSION\020\000\022", - "\034\n\030DELETE_MULTIPLE_VERSIONS\020\001\022\021\n\rDELETE_" + - "FAMILY\020\002\022\031\n\025DELETE_FAMILY_VERSION\020\003\"r\n\rM" + - "utateRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpe" + - "cifier\022 \n\010mutation\030\002 \002(\0132\016.MutationProto" + - "\022\035\n\tcondition\030\003 \001(\0132\n.Condition\"<\n\016Mutat" + - "eResponse\022\027\n\006result\030\001 \001(\0132\007.Result\022\021\n\tpr" + - "ocessed\030\002 \001(\010\"\362\002\n\004Scan\022\027\n\006column\030\001 \003(\0132\007" + - ".Column\022!\n\tattribute\030\002 \003(\0132\016.NameBytesPa" + - "ir\022\020\n\010startRow\030\003 \001(\014\022\017\n\007stopRow\030\004 \001(\014\022\027\n" + - "\006filter\030\005 \001(\0132\007.Filter\022\035\n\ttimeRange\030\006 \001(", - "\0132\n.TimeRange\022\026\n\013maxVersions\030\007 \001(\r:\0011\022\031\n" + - "\013cacheBlocks\030\010 \001(\010:\004true\022\021\n\tbatchSize\030\t " + - "\001(\r\022\025\n\rmaxResultSize\030\n \001(\004\022\022\n\nstoreLimit" + - "\030\013 \001(\r\022\023\n\013storeOffset\030\014 \001(\r\022\"\n\032loadColum" + - "nFamiliesOnDemand\030\r \001(\010\022\024\n\014cachingCount\030" + - "\016 \001(\r\022\023\n\013prefetching\030\017 \001(\010\"\230\001\n\013ScanReque" + - "st\022 \n\006region\030\001 \001(\0132\020.RegionSpecifier\022\023\n\004" + - "scan\030\002 \001(\0132\005.Scan\022\021\n\tscannerId\030\003 \001(\004\022\024\n\014" + - "numberOfRows\030\004 \001(\r\022\024\n\014closeScanner\030\005 \001(\010" + - "\022\023\n\013nextCallSeq\030\006 \001(\004\"l\n\014ScanResponse\022\'\n", - "\016resultCellMeta\030\001 \001(\0132\017.ResultCellMeta\022\021" + - "\n\tscannerId\030\002 \001(\004\022\023\n\013moreResults\030\003 \001(\010\022\013" + - "\n\003ttl\030\004 \001(\r\"%\n\016ResultCellMeta\022\023\n\013cellsLe" + - "ngth\030\001 \003(\r\"\260\001\n\024BulkLoadHFileRequest\022 \n\006r" + - "egion\030\001 \002(\0132\020.RegionSpecifier\0224\n\nfamilyP" + - "ath\030\002 \003(\0132 .BulkLoadHFileRequest.FamilyP" + - "ath\022\024\n\014assignSeqNum\030\003 \001(\010\032*\n\nFamilyPath\022" + - "\016\n\006family\030\001 \002(\014\022\014\n\004path\030\002 \002(\t\"\'\n\025BulkLoa" + - "dHFileResponse\022\016\n\006loaded\030\001 \002(\010\"_\n\026Coproc" + - "essorServiceCall\022\013\n\003row\030\001 \002(\014\022\023\n\013service", - "Name\030\002 \002(\t\022\022\n\nmethodName\030\003 \002(\t\022\017\n\007reques" + - "t\030\004 \002(\014\"d\n\031CoprocessorServiceRequest\022 \n\006" + - "region\030\001 \002(\0132\020.RegionSpecifier\022%\n\004call\030\002" + - " \002(\0132\027.CoprocessorServiceCall\"]\n\032Coproce" + - "ssorServiceResponse\022 \n\006region\030\001 \002(\0132\020.Re" + - "gionSpecifier\022\035\n\005value\030\002 \002(\0132\016.NameBytes" + - "Pair\"B\n\013MultiAction\022 \n\010mutation\030\001 \001(\0132\016." + - "MutationProto\022\021\n\003get\030\002 \001(\0132\004.Get\"I\n\014Acti" + - "onResult\022\026\n\005value\030\001 \001(\0132\007.Result\022!\n\texce" + - "ption\030\002 \001(\0132\016.NameBytesPair\"^\n\014MultiRequ", - "est\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\022\034\n" + - "\006action\030\002 \003(\0132\014.MultiAction\022\016\n\006atomic\030\003 " + - "\001(\010\".\n\rMultiResponse\022\035\n\006result\030\001 \003(\0132\r.A" + - "ctionResult2\342\002\n\rClientService\022 \n\003get\022\013.G" + - "etRequest\032\014.GetResponse\022/\n\010multiGet\022\020.Mu" + - "ltiGetRequest\032\021.MultiGetResponse\022)\n\006muta" + - "te\022\016.MutateRequest\032\017.MutateResponse\022#\n\004s" + - "can\022\014.ScanRequest\032\r.ScanResponse\022>\n\rbulk" + - "LoadHFile\022\025.BulkLoadHFileRequest\032\026.BulkL" + - "oadHFileResponse\022F\n\013execService\022\032.Coproc", - "essorServiceRequest\032\033.CoprocessorService" + - "Response\022&\n\005multi\022\r.MultiRequest\032\016.Multi" + - "ResponseBB\n*org.apache.hadoop.hbase.prot" + - "obuf.generatedB\014ClientProtosH\001\210\001\001\240\001\001" + "\006filter\030\004 \001(\0132\007.Filter\022\036\n\ntime_range\030\005 \001" + + "(\0132\n.TimeRange\022\027\n\014max_versions\030\006 \001(\r:\0011\022" + + "\032\n\014cache_blocks\030\007 \001(\010:\004true\022\023\n\013store_lim" + + "it\030\010 \001(\r\022\024\n\014store_offset\030\t \001(\r\"<\n\006Result" + + "\022\023\n\004cell\030\001 \003(\0132\005.Cell\022\035\n\025associated_cell", + "_count\030\002 \001(\005\"u\n\nGetRequest\022 \n\006region\030\001 \002" + + "(\0132\020.RegionSpecifier\022\021\n\003get\030\002 \002(\0132\004.Get\022" + + "\032\n\022closest_row_before\030\003 \001(\010\022\026\n\016existence" + + "_only\030\004 \001(\010\"z\n\017MultiGetRequest\022 \n\006region" + + "\030\001 \002(\0132\020.RegionSpecifier\022\021\n\003get\030\002 \003(\0132\004." + + "Get\022\032\n\022closest_row_before\030\003 \001(\010\022\026\n\016exist" + + "ence_only\030\004 \001(\010\"6\n\013GetResponse\022\027\n\006result" + + "\030\001 \001(\0132\007.Result\022\016\n\006exists\030\002 \001(\010\";\n\020Multi" + + "GetResponse\022\027\n\006result\030\001 \003(\0132\007.Result\022\016\n\006" + + "exists\030\002 \003(\010\"\200\001\n\tCondition\022\013\n\003row\030\001 \002(\014\022", + "\016\n\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\"\n\014co" + + "mpare_type\030\004 \002(\0162\014.CompareType\022\037\n\ncompar" + + "ator\030\005 \002(\0132\013.Comparator\"\227\006\n\rMutationProt" + + "o\022\013\n\003row\030\001 \001(\014\0220\n\013mutate_type\030\002 \001(\0162\033.Mu" + + "tationProto.MutationType\0220\n\014column_value" + + "\030\003 \003(\0132\032.MutationProto.ColumnValue\022\021\n\tti" + + "mestamp\030\004 \001(\004\022!\n\tattribute\030\005 \003(\0132\016.NameB" + + "ytesPair\022:\n\ndurability\030\006 \001(\0162\031.MutationP" + + "roto.Durability:\013USE_DEFAULT\022\036\n\ntime_ran" + + "ge\030\007 \001(\0132\n.TimeRange\022\035\n\025associated_cell_", + "count\030\010 \001(\005\032\330\001\n\013ColumnValue\022\016\n\006family\030\001 " + + "\002(\014\022B\n\017qualifier_value\030\002 \003(\0132).MutationP" + + "roto.ColumnValue.QualifierValue\032u\n\016Quali" + + "fierValue\022\021\n\tqualifier\030\001 \001(\014\022\r\n\005value\030\002 " + + "\001(\014\022\021\n\ttimestamp\030\003 \001(\004\022.\n\013delete_type\030\004 " + + "\001(\0162\031.MutationProto.DeleteType\"W\n\nDurabi" + + "lity\022\017\n\013USE_DEFAULT\020\000\022\014\n\010SKIP_WAL\020\001\022\r\n\tA" + + "SYNC_WAL\020\002\022\014\n\010SYNC_WAL\020\003\022\r\n\tFSYNC_WAL\020\004\"" + + ">\n\014MutationType\022\n\n\006APPEND\020\000\022\r\n\tINCREMENT" + + "\020\001\022\007\n\003PUT\020\002\022\n\n\006DELETE\020\003\"p\n\nDeleteType\022\026\n", + "\022DELETE_ONE_VERSION\020\000\022\034\n\030DELETE_MULTIPLE" + + "_VERSIONS\020\001\022\021\n\rDELETE_FAMILY\020\002\022\031\n\025DELETE" + + "_FAMILY_VERSION\020\003\"r\n\rMutateRequest\022 \n\006re" + + "gion\030\001 \002(\0132\020.RegionSpecifier\022 \n\010mutation" + + "\030\002 \002(\0132\016.MutationProto\022\035\n\tcondition\030\003 \001(" + + "\0132\n.Condition\"<\n\016MutateResponse\022\027\n\006resul" + + "t\030\001 \001(\0132\007.Result\022\021\n\tprocessed\030\002 \001(\010\"\201\003\n\004" + + "Scan\022\027\n\006column\030\001 \003(\0132\007.Column\022!\n\tattribu" + + "te\030\002 \003(\0132\016.NameBytesPair\022\021\n\tstart_row\030\003 " + + "\001(\014\022\020\n\010stop_row\030\004 \001(\014\022\027\n\006filter\030\005 \001(\0132\007.", + "Filter\022\036\n\ntime_range\030\006 \001(\0132\n.TimeRange\022\027" + + "\n\014max_versions\030\007 \001(\r:\0011\022\032\n\014cache_blocks\030" + + "\010 \001(\010:\004true\022\022\n\nbatch_size\030\t \001(\r\022\027\n\017max_r" + + "esult_size\030\n \001(\004\022\023\n\013store_limit\030\013 \001(\r\022\024\n" + + "\014store_offset\030\014 \001(\r\022&\n\036load_column_famil" + + "ies_on_demand\030\r \001(\010\022\025\n\rcaching_count\030\016 \001" + + "(\r\022\023\n\013prefetching\030\017 \001(\010\"\236\001\n\013ScanRequest\022" + + " \n\006region\030\001 \001(\0132\020.RegionSpecifier\022\023\n\004sca" + + "n\030\002 \001(\0132\005.Scan\022\022\n\nscanner_id\030\003 \001(\004\022\026\n\016nu" + + "mber_of_rows\030\004 \001(\r\022\025\n\rclose_scanner\030\005 \001(", + "\010\022\025\n\rnext_call_seq\030\006 \001(\004\"p\n\014ScanResponse" + + "\022)\n\020result_cell_meta\030\001 \001(\0132\017.ResultCellM" + + "eta\022\022\n\nscanner_id\030\002 \001(\004\022\024\n\014more_results\030" + + "\003 \001(\010\022\013\n\003ttl\030\004 \001(\r\"&\n\016ResultCellMeta\022\024\n\014" + + "cells_length\030\001 \003(\r\"\263\001\n\024BulkLoadHFileRequ" + + "est\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\0225\n" + + "\013family_path\030\002 \003(\0132 .BulkLoadHFileReques" + + "t.FamilyPath\022\026\n\016assign_seq_num\030\003 \001(\010\032*\n\n" + + "FamilyPath\022\016\n\006family\030\001 \002(\014\022\014\n\004path\030\002 \002(\t" + + "\"\'\n\025BulkLoadHFileResponse\022\016\n\006loaded\030\001 \002(", + "\010\"a\n\026CoprocessorServiceCall\022\013\n\003row\030\001 \002(\014" + + "\022\024\n\014service_name\030\002 \002(\t\022\023\n\013method_name\030\003 " + + "\002(\t\022\017\n\007request\030\004 \002(\014\"d\n\031CoprocessorServi" + + "ceRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpecif" + + "ier\022%\n\004call\030\002 \002(\0132\027.CoprocessorServiceCa" + + "ll\"]\n\032CoprocessorServiceResponse\022 \n\006regi" + + "on\030\001 \002(\0132\020.RegionSpecifier\022\035\n\005value\030\002 \002(" + + "\0132\016.NameBytesPair\"B\n\013MultiAction\022 \n\010muta" + + "tion\030\001 \001(\0132\016.MutationProto\022\021\n\003get\030\002 \001(\0132" + + "\004.Get\"I\n\014ActionResult\022\026\n\005value\030\001 \001(\0132\007.R", + "esult\022!\n\texception\030\002 \001(\0132\016.NameBytesPair" + + "\"^\n\014MultiRequest\022 \n\006region\030\001 \002(\0132\020.Regio" + + "nSpecifier\022\034\n\006action\030\002 \003(\0132\014.MultiAction" + + "\022\016\n\006atomic\030\003 \001(\010\".\n\rMultiResponse\022\035\n\006res" + + "ult\030\001 \003(\0132\r.ActionResult2\342\002\n\rClientServi" + + "ce\022 \n\003Get\022\013.GetRequest\032\014.GetResponse\022/\n\010" + + "MultiGet\022\020.MultiGetRequest\032\021.MultiGetRes" + + "ponse\022)\n\006Mutate\022\016.MutateRequest\032\017.Mutate" + + "Response\022#\n\004Scan\022\014.ScanRequest\032\r.ScanRes" + + "ponse\022>\n\rBulkLoadHFile\022\025.BulkLoadHFileRe", + "quest\032\026.BulkLoadHFileResponse\022F\n\013ExecSer" + + "vice\022\032.CoprocessorServiceRequest\032\033.Copro" + + "cessorServiceResponse\022&\n\005Multi\022\r.MultiRe" + + "quest\032\016.MultiResponseBB\n*org.apache.hado" + + "op.hbase.protobuf.generatedB\014ClientProto" + + "sH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterIdProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterIdProtos.java index aac3b80..c9476c4 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterIdProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterIdProtos.java @@ -11,7 +11,7 @@ public final class ClusterIdProtos { public interface ClusterIdOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string clusterId = 1; + // required string cluster_id = 1; boolean hasClusterId(); String getClusterId(); } @@ -44,8 +44,8 @@ public final class ClusterIdProtos { } private int bitField0_; - // required string clusterId = 1; - public static final int CLUSTERID_FIELD_NUMBER = 1; + // required string cluster_id = 1; + public static final int CLUSTER_ID_FIELD_NUMBER = 1; private java.lang.Object clusterId_; public boolean hasClusterId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -149,7 +149,7 @@ public final class ClusterIdProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasClusterId()) { - hash = (37 * hash) + CLUSTERID_FIELD_NUMBER; + hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER; hash = (53 * hash) + getClusterId().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -377,7 +377,7 @@ public final class ClusterIdProtos { private int bitField0_; - // required string clusterId = 1; + // required string cluster_id = 1; private java.lang.Object clusterId_ = ""; public boolean hasClusterId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -438,9 +438,9 @@ public final class ClusterIdProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\017ClusterId.proto\"\036\n\tClusterId\022\021\n\tcluste" + - "rId\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pro" + - "tobuf.generatedB\017ClusterIdProtosH\001\240\001\001" + "\n\017ClusterId.proto\"\037\n\tClusterId\022\022\n\ncluste" + + "r_id\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pr" + + "otobuf.generatedB\017ClusterIdProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java index 3961689..2500610 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClusterStatusProtos.java @@ -11,7 +11,7 @@ public final class ClusterStatusProtos { public interface RegionStateOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .RegionInfo regionInfo = 1; + // required .RegionInfo region_info = 1; boolean hasRegionInfo(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(); @@ -155,8 +155,8 @@ public final class ClusterStatusProtos { } private int bitField0_; - // required .RegionInfo regionInfo = 1; - public static final int REGIONINFO_FIELD_NUMBER = 1; + // required .RegionInfo region_info = 1; + public static final int REGION_INFO_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_; public boolean hasRegionInfo() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -295,7 +295,7 @@ public final class ClusterStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionInfo()) { - hash = (37 * hash) + REGIONINFO_FIELD_NUMBER; + hash = (37 * hash) + REGION_INFO_FIELD_NUMBER; hash = (53 * hash) + getRegionInfo().hashCode(); } if (hasState()) { @@ -586,7 +586,7 @@ public final class ClusterStatusProtos { private int bitField0_; - // required .RegionInfo regionInfo = 1; + // required .RegionInfo region_info = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_; @@ -740,7 +740,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getSpec(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getSpecOrBuilder(); - // required .RegionState regionState = 2; + // required .RegionState region_state = 2; boolean hasRegionState(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState getRegionState(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionStateOrBuilder getRegionStateOrBuilder(); @@ -787,8 +787,8 @@ public final class ClusterStatusProtos { return spec_; } - // required .RegionState regionState = 2; - public static final int REGIONSTATE_FIELD_NUMBER = 2; + // required .RegionState region_state = 2; + public static final int REGION_STATE_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState regionState_; public boolean hasRegionState() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -902,7 +902,7 @@ public final class ClusterStatusProtos { hash = (53 * hash) + getSpec().hashCode(); } if (hasRegionState()) { - hash = (37 * hash) + REGIONSTATE_FIELD_NUMBER; + hash = (37 * hash) + REGION_STATE_FIELD_NUMBER; hash = (53 * hash) + getRegionState().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1272,7 +1272,7 @@ public final class ClusterStatusProtos { return specBuilder_; } - // required .RegionState regionState = 2; + // required .RegionState region_state = 2; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState regionState_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionStateOrBuilder> regionStateBuilder_; @@ -1376,7 +1376,7 @@ public final class ClusterStatusProtos { public interface RegionLoadOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .RegionSpecifier regionSpecifier = 1; + // required .RegionSpecifier region_specifier = 1; boolean hasRegionSpecifier(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder(); @@ -1389,51 +1389,51 @@ public final class ClusterStatusProtos { boolean hasStorefiles(); int getStorefiles(); - // optional uint32 storeUncompressedSizeMB = 4; + // optional uint32 store_uncompressed_size_MB = 4; boolean hasStoreUncompressedSizeMB(); int getStoreUncompressedSizeMB(); - // optional uint32 storefileSizeMB = 5; + // optional uint32 storefile_size_MB = 5; boolean hasStorefileSizeMB(); int getStorefileSizeMB(); - // optional uint32 memstoreSizeMB = 6; + // optional uint32 memstore_size_MB = 6; boolean hasMemstoreSizeMB(); int getMemstoreSizeMB(); - // optional uint32 storefileIndexSizeMB = 7; + // optional uint32 storefile_index_size_MB = 7; boolean hasStorefileIndexSizeMB(); int getStorefileIndexSizeMB(); - // optional uint64 readRequestsCount = 8; + // optional uint64 read_requests_count = 8; boolean hasReadRequestsCount(); long getReadRequestsCount(); - // optional uint64 writeRequestsCount = 9; + // optional uint64 write_requests_count = 9; boolean hasWriteRequestsCount(); long getWriteRequestsCount(); - // optional uint64 totalCompactingKVs = 10; + // optional uint64 total_compacting_KVs = 10; boolean hasTotalCompactingKVs(); long getTotalCompactingKVs(); - // optional uint64 currentCompactedKVs = 11; + // optional uint64 current_compacted_KVs = 11; boolean hasCurrentCompactedKVs(); long getCurrentCompactedKVs(); - // optional uint32 rootIndexSizeKB = 12; + // optional uint32 root_index_size_KB = 12; boolean hasRootIndexSizeKB(); int getRootIndexSizeKB(); - // optional uint32 totalStaticIndexSizeKB = 13; + // optional uint32 total_static_index_size_KB = 13; boolean hasTotalStaticIndexSizeKB(); int getTotalStaticIndexSizeKB(); - // optional uint32 totalStaticBloomSizeKB = 14; + // optional uint32 total_static_bloom_size_KB = 14; boolean hasTotalStaticBloomSizeKB(); int getTotalStaticBloomSizeKB(); - // optional uint64 completeSequenceId = 15; + // optional uint64 complete_sequence_id = 15; boolean hasCompleteSequenceId(); long getCompleteSequenceId(); } @@ -1466,8 +1466,8 @@ public final class ClusterStatusProtos { } private int bitField0_; - // required .RegionSpecifier regionSpecifier = 1; - public static final int REGIONSPECIFIER_FIELD_NUMBER = 1; + // required .RegionSpecifier region_specifier = 1; + public static final int REGION_SPECIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_; public boolean hasRegionSpecifier() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1499,8 +1499,8 @@ public final class ClusterStatusProtos { return storefiles_; } - // optional uint32 storeUncompressedSizeMB = 4; - public static final int STOREUNCOMPRESSEDSIZEMB_FIELD_NUMBER = 4; + // optional uint32 store_uncompressed_size_MB = 4; + public static final int STORE_UNCOMPRESSED_SIZE_MB_FIELD_NUMBER = 4; private int storeUncompressedSizeMB_; public boolean hasStoreUncompressedSizeMB() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1509,8 +1509,8 @@ public final class ClusterStatusProtos { return storeUncompressedSizeMB_; } - // optional uint32 storefileSizeMB = 5; - public static final int STOREFILESIZEMB_FIELD_NUMBER = 5; + // optional uint32 storefile_size_MB = 5; + public static final int STOREFILE_SIZE_MB_FIELD_NUMBER = 5; private int storefileSizeMB_; public boolean hasStorefileSizeMB() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -1519,8 +1519,8 @@ public final class ClusterStatusProtos { return storefileSizeMB_; } - // optional uint32 memstoreSizeMB = 6; - public static final int MEMSTORESIZEMB_FIELD_NUMBER = 6; + // optional uint32 memstore_size_MB = 6; + public static final int MEMSTORE_SIZE_MB_FIELD_NUMBER = 6; private int memstoreSizeMB_; public boolean hasMemstoreSizeMB() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -1529,8 +1529,8 @@ public final class ClusterStatusProtos { return memstoreSizeMB_; } - // optional uint32 storefileIndexSizeMB = 7; - public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 7; + // optional uint32 storefile_index_size_MB = 7; + public static final int STOREFILE_INDEX_SIZE_MB_FIELD_NUMBER = 7; private int storefileIndexSizeMB_; public boolean hasStorefileIndexSizeMB() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -1539,8 +1539,8 @@ public final class ClusterStatusProtos { return storefileIndexSizeMB_; } - // optional uint64 readRequestsCount = 8; - public static final int READREQUESTSCOUNT_FIELD_NUMBER = 8; + // optional uint64 read_requests_count = 8; + public static final int READ_REQUESTS_COUNT_FIELD_NUMBER = 8; private long readRequestsCount_; public boolean hasReadRequestsCount() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -1549,8 +1549,8 @@ public final class ClusterStatusProtos { return readRequestsCount_; } - // optional uint64 writeRequestsCount = 9; - public static final int WRITEREQUESTSCOUNT_FIELD_NUMBER = 9; + // optional uint64 write_requests_count = 9; + public static final int WRITE_REQUESTS_COUNT_FIELD_NUMBER = 9; private long writeRequestsCount_; public boolean hasWriteRequestsCount() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -1559,8 +1559,8 @@ public final class ClusterStatusProtos { return writeRequestsCount_; } - // optional uint64 totalCompactingKVs = 10; - public static final int TOTALCOMPACTINGKVS_FIELD_NUMBER = 10; + // optional uint64 total_compacting_KVs = 10; + public static final int TOTAL_COMPACTING_KVS_FIELD_NUMBER = 10; private long totalCompactingKVs_; public boolean hasTotalCompactingKVs() { return ((bitField0_ & 0x00000200) == 0x00000200); @@ -1569,8 +1569,8 @@ public final class ClusterStatusProtos { return totalCompactingKVs_; } - // optional uint64 currentCompactedKVs = 11; - public static final int CURRENTCOMPACTEDKVS_FIELD_NUMBER = 11; + // optional uint64 current_compacted_KVs = 11; + public static final int CURRENT_COMPACTED_KVS_FIELD_NUMBER = 11; private long currentCompactedKVs_; public boolean hasCurrentCompactedKVs() { return ((bitField0_ & 0x00000400) == 0x00000400); @@ -1579,8 +1579,8 @@ public final class ClusterStatusProtos { return currentCompactedKVs_; } - // optional uint32 rootIndexSizeKB = 12; - public static final int ROOTINDEXSIZEKB_FIELD_NUMBER = 12; + // optional uint32 root_index_size_KB = 12; + public static final int ROOT_INDEX_SIZE_KB_FIELD_NUMBER = 12; private int rootIndexSizeKB_; public boolean hasRootIndexSizeKB() { return ((bitField0_ & 0x00000800) == 0x00000800); @@ -1589,8 +1589,8 @@ public final class ClusterStatusProtos { return rootIndexSizeKB_; } - // optional uint32 totalStaticIndexSizeKB = 13; - public static final int TOTALSTATICINDEXSIZEKB_FIELD_NUMBER = 13; + // optional uint32 total_static_index_size_KB = 13; + public static final int TOTAL_STATIC_INDEX_SIZE_KB_FIELD_NUMBER = 13; private int totalStaticIndexSizeKB_; public boolean hasTotalStaticIndexSizeKB() { return ((bitField0_ & 0x00001000) == 0x00001000); @@ -1599,8 +1599,8 @@ public final class ClusterStatusProtos { return totalStaticIndexSizeKB_; } - // optional uint32 totalStaticBloomSizeKB = 14; - public static final int TOTALSTATICBLOOMSIZEKB_FIELD_NUMBER = 14; + // optional uint32 total_static_bloom_size_KB = 14; + public static final int TOTAL_STATIC_BLOOM_SIZE_KB_FIELD_NUMBER = 14; private int totalStaticBloomSizeKB_; public boolean hasTotalStaticBloomSizeKB() { return ((bitField0_ & 0x00002000) == 0x00002000); @@ -1609,8 +1609,8 @@ public final class ClusterStatusProtos { return totalStaticBloomSizeKB_; } - // optional uint64 completeSequenceId = 15; - public static final int COMPLETESEQUENCEID_FIELD_NUMBER = 15; + // optional uint64 complete_sequence_id = 15; + public static final int COMPLETE_SEQUENCE_ID_FIELD_NUMBER = 15; private long completeSequenceId_; public boolean hasCompleteSequenceId() { return ((bitField0_ & 0x00004000) == 0x00004000); @@ -1878,7 +1878,7 @@ public final class ClusterStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionSpecifier()) { - hash = (37 * hash) + REGIONSPECIFIER_FIELD_NUMBER; + hash = (37 * hash) + REGION_SPECIFIER_FIELD_NUMBER; hash = (53 * hash) + getRegionSpecifier().hashCode(); } if (hasStores()) { @@ -1890,51 +1890,51 @@ public final class ClusterStatusProtos { hash = (53 * hash) + getStorefiles(); } if (hasStoreUncompressedSizeMB()) { - hash = (37 * hash) + STOREUNCOMPRESSEDSIZEMB_FIELD_NUMBER; + hash = (37 * hash) + STORE_UNCOMPRESSED_SIZE_MB_FIELD_NUMBER; hash = (53 * hash) + getStoreUncompressedSizeMB(); } if (hasStorefileSizeMB()) { - hash = (37 * hash) + STOREFILESIZEMB_FIELD_NUMBER; + hash = (37 * hash) + STOREFILE_SIZE_MB_FIELD_NUMBER; hash = (53 * hash) + getStorefileSizeMB(); } if (hasMemstoreSizeMB()) { - hash = (37 * hash) + MEMSTORESIZEMB_FIELD_NUMBER; + hash = (37 * hash) + MEMSTORE_SIZE_MB_FIELD_NUMBER; hash = (53 * hash) + getMemstoreSizeMB(); } if (hasStorefileIndexSizeMB()) { - hash = (37 * hash) + STOREFILEINDEXSIZEMB_FIELD_NUMBER; + hash = (37 * hash) + STOREFILE_INDEX_SIZE_MB_FIELD_NUMBER; hash = (53 * hash) + getStorefileIndexSizeMB(); } if (hasReadRequestsCount()) { - hash = (37 * hash) + READREQUESTSCOUNT_FIELD_NUMBER; + hash = (37 * hash) + READ_REQUESTS_COUNT_FIELD_NUMBER; hash = (53 * hash) + hashLong(getReadRequestsCount()); } if (hasWriteRequestsCount()) { - hash = (37 * hash) + WRITEREQUESTSCOUNT_FIELD_NUMBER; + hash = (37 * hash) + WRITE_REQUESTS_COUNT_FIELD_NUMBER; hash = (53 * hash) + hashLong(getWriteRequestsCount()); } if (hasTotalCompactingKVs()) { - hash = (37 * hash) + TOTALCOMPACTINGKVS_FIELD_NUMBER; + hash = (37 * hash) + TOTAL_COMPACTING_KVS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getTotalCompactingKVs()); } if (hasCurrentCompactedKVs()) { - hash = (37 * hash) + CURRENTCOMPACTEDKVS_FIELD_NUMBER; + hash = (37 * hash) + CURRENT_COMPACTED_KVS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCurrentCompactedKVs()); } if (hasRootIndexSizeKB()) { - hash = (37 * hash) + ROOTINDEXSIZEKB_FIELD_NUMBER; + hash = (37 * hash) + ROOT_INDEX_SIZE_KB_FIELD_NUMBER; hash = (53 * hash) + getRootIndexSizeKB(); } if (hasTotalStaticIndexSizeKB()) { - hash = (37 * hash) + TOTALSTATICINDEXSIZEKB_FIELD_NUMBER; + hash = (37 * hash) + TOTAL_STATIC_INDEX_SIZE_KB_FIELD_NUMBER; hash = (53 * hash) + getTotalStaticIndexSizeKB(); } if (hasTotalStaticBloomSizeKB()) { - hash = (37 * hash) + TOTALSTATICBLOOMSIZEKB_FIELD_NUMBER; + hash = (37 * hash) + TOTAL_STATIC_BLOOM_SIZE_KB_FIELD_NUMBER; hash = (53 * hash) + getTotalStaticBloomSizeKB(); } if (hasCompleteSequenceId()) { - hash = (37 * hash) + COMPLETESEQUENCEID_FIELD_NUMBER; + hash = (37 * hash) + COMPLETE_SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCompleteSequenceId()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2375,7 +2375,7 @@ public final class ClusterStatusProtos { private int bitField0_; - // required .RegionSpecifier regionSpecifier = 1; + // required .RegionSpecifier region_specifier = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionSpecifierBuilder_; @@ -2507,7 +2507,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 storeUncompressedSizeMB = 4; + // optional uint32 store_uncompressed_size_MB = 4; private int storeUncompressedSizeMB_ ; public boolean hasStoreUncompressedSizeMB() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2528,7 +2528,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 storefileSizeMB = 5; + // optional uint32 storefile_size_MB = 5; private int storefileSizeMB_ ; public boolean hasStorefileSizeMB() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -2549,7 +2549,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 memstoreSizeMB = 6; + // optional uint32 memstore_size_MB = 6; private int memstoreSizeMB_ ; public boolean hasMemstoreSizeMB() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -2570,7 +2570,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 storefileIndexSizeMB = 7; + // optional uint32 storefile_index_size_MB = 7; private int storefileIndexSizeMB_ ; public boolean hasStorefileIndexSizeMB() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -2591,7 +2591,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint64 readRequestsCount = 8; + // optional uint64 read_requests_count = 8; private long readRequestsCount_ ; public boolean hasReadRequestsCount() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -2612,7 +2612,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint64 writeRequestsCount = 9; + // optional uint64 write_requests_count = 9; private long writeRequestsCount_ ; public boolean hasWriteRequestsCount() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -2633,7 +2633,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint64 totalCompactingKVs = 10; + // optional uint64 total_compacting_KVs = 10; private long totalCompactingKVs_ ; public boolean hasTotalCompactingKVs() { return ((bitField0_ & 0x00000200) == 0x00000200); @@ -2654,7 +2654,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint64 currentCompactedKVs = 11; + // optional uint64 current_compacted_KVs = 11; private long currentCompactedKVs_ ; public boolean hasCurrentCompactedKVs() { return ((bitField0_ & 0x00000400) == 0x00000400); @@ -2675,7 +2675,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 rootIndexSizeKB = 12; + // optional uint32 root_index_size_KB = 12; private int rootIndexSizeKB_ ; public boolean hasRootIndexSizeKB() { return ((bitField0_ & 0x00000800) == 0x00000800); @@ -2696,7 +2696,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 totalStaticIndexSizeKB = 13; + // optional uint32 total_static_index_size_KB = 13; private int totalStaticIndexSizeKB_ ; public boolean hasTotalStaticIndexSizeKB() { return ((bitField0_ & 0x00001000) == 0x00001000); @@ -2717,7 +2717,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 totalStaticBloomSizeKB = 14; + // optional uint32 total_static_bloom_size_KB = 14; private int totalStaticBloomSizeKB_ ; public boolean hasTotalStaticBloomSizeKB() { return ((bitField0_ & 0x00002000) == 0x00002000); @@ -2738,7 +2738,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint64 completeSequenceId = 15; + // optional uint64 complete_sequence_id = 15; private long completeSequenceId_ ; public boolean hasCompleteSequenceId() { return ((bitField0_ & 0x00004000) == 0x00004000); @@ -2773,23 +2773,23 @@ public final class ClusterStatusProtos { public interface ServerLoadOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint32 numberOfRequests = 1; + // optional uint32 number_of_requests = 1; boolean hasNumberOfRequests(); int getNumberOfRequests(); - // optional uint32 totalNumberOfRequests = 2; + // optional uint32 total_number_of_requests = 2; boolean hasTotalNumberOfRequests(); int getTotalNumberOfRequests(); - // optional uint32 usedHeapMB = 3; + // optional uint32 used_heap_MB = 3; boolean hasUsedHeapMB(); int getUsedHeapMB(); - // optional uint32 maxHeapMB = 4; + // optional uint32 max_heap_MB = 4; boolean hasMaxHeapMB(); int getMaxHeapMB(); - // repeated .RegionLoad regionLoads = 5; + // repeated .RegionLoad region_loads = 5; java.util.List getRegionLoadsList(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad getRegionLoads(int index); @@ -2809,15 +2809,15 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CoprocessorOrBuilder getCoprocessorsOrBuilder( int index); - // optional uint64 reportStartTime = 7; + // optional uint64 report_start_time = 7; boolean hasReportStartTime(); long getReportStartTime(); - // optional uint64 reportEndTime = 8; + // optional uint64 report_end_time = 8; boolean hasReportEndTime(); long getReportEndTime(); - // optional uint32 infoServerPort = 9; + // optional uint32 info_server_port = 9; boolean hasInfoServerPort(); int getInfoServerPort(); } @@ -2850,8 +2850,8 @@ public final class ClusterStatusProtos { } private int bitField0_; - // optional uint32 numberOfRequests = 1; - public static final int NUMBEROFREQUESTS_FIELD_NUMBER = 1; + // optional uint32 number_of_requests = 1; + public static final int NUMBER_OF_REQUESTS_FIELD_NUMBER = 1; private int numberOfRequests_; public boolean hasNumberOfRequests() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2860,8 +2860,8 @@ public final class ClusterStatusProtos { return numberOfRequests_; } - // optional uint32 totalNumberOfRequests = 2; - public static final int TOTALNUMBEROFREQUESTS_FIELD_NUMBER = 2; + // optional uint32 total_number_of_requests = 2; + public static final int TOTAL_NUMBER_OF_REQUESTS_FIELD_NUMBER = 2; private int totalNumberOfRequests_; public boolean hasTotalNumberOfRequests() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2870,8 +2870,8 @@ public final class ClusterStatusProtos { return totalNumberOfRequests_; } - // optional uint32 usedHeapMB = 3; - public static final int USEDHEAPMB_FIELD_NUMBER = 3; + // optional uint32 used_heap_MB = 3; + public static final int USED_HEAP_MB_FIELD_NUMBER = 3; private int usedHeapMB_; public boolean hasUsedHeapMB() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2880,8 +2880,8 @@ public final class ClusterStatusProtos { return usedHeapMB_; } - // optional uint32 maxHeapMB = 4; - public static final int MAXHEAPMB_FIELD_NUMBER = 4; + // optional uint32 max_heap_MB = 4; + public static final int MAX_HEAP_MB_FIELD_NUMBER = 4; private int maxHeapMB_; public boolean hasMaxHeapMB() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2890,8 +2890,8 @@ public final class ClusterStatusProtos { return maxHeapMB_; } - // repeated .RegionLoad regionLoads = 5; - public static final int REGIONLOADS_FIELD_NUMBER = 5; + // repeated .RegionLoad region_loads = 5; + public static final int REGION_LOADS_FIELD_NUMBER = 5; private java.util.List regionLoads_; public java.util.List getRegionLoadsList() { return regionLoads_; @@ -2932,8 +2932,8 @@ public final class ClusterStatusProtos { return coprocessors_.get(index); } - // optional uint64 reportStartTime = 7; - public static final int REPORTSTARTTIME_FIELD_NUMBER = 7; + // optional uint64 report_start_time = 7; + public static final int REPORT_START_TIME_FIELD_NUMBER = 7; private long reportStartTime_; public boolean hasReportStartTime() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -2942,8 +2942,8 @@ public final class ClusterStatusProtos { return reportStartTime_; } - // optional uint64 reportEndTime = 8; - public static final int REPORTENDTIME_FIELD_NUMBER = 8; + // optional uint64 report_end_time = 8; + public static final int REPORT_END_TIME_FIELD_NUMBER = 8; private long reportEndTime_; public boolean hasReportEndTime() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -2952,8 +2952,8 @@ public final class ClusterStatusProtos { return reportEndTime_; } - // optional uint32 infoServerPort = 9; - public static final int INFOSERVERPORT_FIELD_NUMBER = 9; + // optional uint32 info_server_port = 9; + public static final int INFO_SERVER_PORT_FIELD_NUMBER = 9; private int infoServerPort_; public boolean hasInfoServerPort() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -3141,23 +3141,23 @@ public final class ClusterStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasNumberOfRequests()) { - hash = (37 * hash) + NUMBEROFREQUESTS_FIELD_NUMBER; + hash = (37 * hash) + NUMBER_OF_REQUESTS_FIELD_NUMBER; hash = (53 * hash) + getNumberOfRequests(); } if (hasTotalNumberOfRequests()) { - hash = (37 * hash) + TOTALNUMBEROFREQUESTS_FIELD_NUMBER; + hash = (37 * hash) + TOTAL_NUMBER_OF_REQUESTS_FIELD_NUMBER; hash = (53 * hash) + getTotalNumberOfRequests(); } if (hasUsedHeapMB()) { - hash = (37 * hash) + USEDHEAPMB_FIELD_NUMBER; + hash = (37 * hash) + USED_HEAP_MB_FIELD_NUMBER; hash = (53 * hash) + getUsedHeapMB(); } if (hasMaxHeapMB()) { - hash = (37 * hash) + MAXHEAPMB_FIELD_NUMBER; + hash = (37 * hash) + MAX_HEAP_MB_FIELD_NUMBER; hash = (53 * hash) + getMaxHeapMB(); } if (getRegionLoadsCount() > 0) { - hash = (37 * hash) + REGIONLOADS_FIELD_NUMBER; + hash = (37 * hash) + REGION_LOADS_FIELD_NUMBER; hash = (53 * hash) + getRegionLoadsList().hashCode(); } if (getCoprocessorsCount() > 0) { @@ -3165,15 +3165,15 @@ public final class ClusterStatusProtos { hash = (53 * hash) + getCoprocessorsList().hashCode(); } if (hasReportStartTime()) { - hash = (37 * hash) + REPORTSTARTTIME_FIELD_NUMBER; + hash = (37 * hash) + REPORT_START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getReportStartTime()); } if (hasReportEndTime()) { - hash = (37 * hash) + REPORTENDTIME_FIELD_NUMBER; + hash = (37 * hash) + REPORT_END_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getReportEndTime()); } if (hasInfoServerPort()) { - hash = (37 * hash) + INFOSERVERPORT_FIELD_NUMBER; + hash = (37 * hash) + INFO_SERVER_PORT_FIELD_NUMBER; hash = (53 * hash) + getInfoServerPort(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3589,7 +3589,7 @@ public final class ClusterStatusProtos { private int bitField0_; - // optional uint32 numberOfRequests = 1; + // optional uint32 number_of_requests = 1; private int numberOfRequests_ ; public boolean hasNumberOfRequests() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3610,7 +3610,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 totalNumberOfRequests = 2; + // optional uint32 total_number_of_requests = 2; private int totalNumberOfRequests_ ; public boolean hasTotalNumberOfRequests() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3631,7 +3631,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 usedHeapMB = 3; + // optional uint32 used_heap_MB = 3; private int usedHeapMB_ ; public boolean hasUsedHeapMB() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -3652,7 +3652,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 maxHeapMB = 4; + // optional uint32 max_heap_MB = 4; private int maxHeapMB_ ; public boolean hasMaxHeapMB() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -3673,7 +3673,7 @@ public final class ClusterStatusProtos { return this; } - // repeated .RegionLoad regionLoads = 5; + // repeated .RegionLoad region_loads = 5; private java.util.List regionLoads_ = java.util.Collections.emptyList(); private void ensureRegionLoadsIsMutable() { @@ -4045,7 +4045,7 @@ public final class ClusterStatusProtos { return coprocessorsBuilder_; } - // optional uint64 reportStartTime = 7; + // optional uint64 report_start_time = 7; private long reportStartTime_ ; public boolean hasReportStartTime() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -4066,7 +4066,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint64 reportEndTime = 8; + // optional uint64 report_end_time = 8; private long reportEndTime_ ; public boolean hasReportEndTime() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -4087,7 +4087,7 @@ public final class ClusterStatusProtos { return this; } - // optional uint32 infoServerPort = 9; + // optional uint32 info_server_port = 9; private int infoServerPort_ ; public boolean hasInfoServerPort() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -4127,7 +4127,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); - // required .ServerLoad serverLoad = 2; + // required .ServerLoad server_load = 2; boolean hasServerLoad(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad getServerLoad(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder getServerLoadOrBuilder(); @@ -4174,8 +4174,8 @@ public final class ClusterStatusProtos { return server_; } - // required .ServerLoad serverLoad = 2; - public static final int SERVERLOAD_FIELD_NUMBER = 2; + // required .ServerLoad server_load = 2; + public static final int SERVER_LOAD_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad serverLoad_; public boolean hasServerLoad() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -4289,7 +4289,7 @@ public final class ClusterStatusProtos { hash = (53 * hash) + getServer().hashCode(); } if (hasServerLoad()) { - hash = (37 * hash) + SERVERLOAD_FIELD_NUMBER; + hash = (37 * hash) + SERVER_LOAD_FIELD_NUMBER; hash = (53 * hash) + getServerLoad().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -4659,7 +4659,7 @@ public final class ClusterStatusProtos { return serverBuilder_; } - // required .ServerLoad serverLoad = 2; + // required .ServerLoad server_load = 2; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad serverLoad_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoad.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ServerLoadOrBuilder> serverLoadBuilder_; @@ -4763,12 +4763,12 @@ public final class ClusterStatusProtos { public interface ClusterStatusOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional .HBaseVersionFileContent hbaseVersion = 1; + // optional .HBaseVersionFileContent hbase_version = 1; boolean hasHbaseVersion(); org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContent getHbaseVersion(); org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder getHbaseVersionOrBuilder(); - // repeated .LiveServerInfo liveServers = 2; + // repeated .LiveServerInfo live_servers = 2; java.util.List getLiveServersList(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.LiveServerInfo getLiveServers(int index); @@ -4778,7 +4778,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.LiveServerInfoOrBuilder getLiveServersOrBuilder( int index); - // repeated .ServerName deadServers = 3; + // repeated .ServerName dead_servers = 3; java.util.List getDeadServersList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDeadServers(int index); @@ -4788,7 +4788,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDeadServersOrBuilder( int index); - // repeated .RegionInTransition regionsInTransition = 4; + // repeated .RegionInTransition regions_in_transition = 4; java.util.List getRegionsInTransitionList(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionInTransition getRegionsInTransition(int index); @@ -4798,12 +4798,12 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionInTransitionOrBuilder getRegionsInTransitionOrBuilder( int index); - // optional .ClusterId clusterId = 5; + // optional .ClusterId cluster_id = 5; boolean hasClusterId(); org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterId getClusterId(); org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterIdOrBuilder getClusterIdOrBuilder(); - // repeated .Coprocessor masterCoprocessors = 6; + // repeated .Coprocessor master_coprocessors = 6; java.util.List getMasterCoprocessorsList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.Coprocessor getMasterCoprocessors(int index); @@ -4818,7 +4818,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder(); - // repeated .ServerName backupMasters = 8; + // repeated .ServerName backup_masters = 8; java.util.List getBackupMastersList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getBackupMasters(int index); @@ -4828,7 +4828,7 @@ public final class ClusterStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getBackupMastersOrBuilder( int index); - // optional bool balancerOn = 9; + // optional bool balancer_on = 9; boolean hasBalancerOn(); boolean getBalancerOn(); } @@ -4861,8 +4861,8 @@ public final class ClusterStatusProtos { } private int bitField0_; - // optional .HBaseVersionFileContent hbaseVersion = 1; - public static final int HBASEVERSION_FIELD_NUMBER = 1; + // optional .HBaseVersionFileContent hbase_version = 1; + public static final int HBASE_VERSION_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContent hbaseVersion_; public boolean hasHbaseVersion() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -4874,8 +4874,8 @@ public final class ClusterStatusProtos { return hbaseVersion_; } - // repeated .LiveServerInfo liveServers = 2; - public static final int LIVESERVERS_FIELD_NUMBER = 2; + // repeated .LiveServerInfo live_servers = 2; + public static final int LIVE_SERVERS_FIELD_NUMBER = 2; private java.util.List liveServers_; public java.util.List getLiveServersList() { return liveServers_; @@ -4895,8 +4895,8 @@ public final class ClusterStatusProtos { return liveServers_.get(index); } - // repeated .ServerName deadServers = 3; - public static final int DEADSERVERS_FIELD_NUMBER = 3; + // repeated .ServerName dead_servers = 3; + public static final int DEAD_SERVERS_FIELD_NUMBER = 3; private java.util.List deadServers_; public java.util.List getDeadServersList() { return deadServers_; @@ -4916,8 +4916,8 @@ public final class ClusterStatusProtos { return deadServers_.get(index); } - // repeated .RegionInTransition regionsInTransition = 4; - public static final int REGIONSINTRANSITION_FIELD_NUMBER = 4; + // repeated .RegionInTransition regions_in_transition = 4; + public static final int REGIONS_IN_TRANSITION_FIELD_NUMBER = 4; private java.util.List regionsInTransition_; public java.util.List getRegionsInTransitionList() { return regionsInTransition_; @@ -4937,8 +4937,8 @@ public final class ClusterStatusProtos { return regionsInTransition_.get(index); } - // optional .ClusterId clusterId = 5; - public static final int CLUSTERID_FIELD_NUMBER = 5; + // optional .ClusterId cluster_id = 5; + public static final int CLUSTER_ID_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterId clusterId_; public boolean hasClusterId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -4950,8 +4950,8 @@ public final class ClusterStatusProtos { return clusterId_; } - // repeated .Coprocessor masterCoprocessors = 6; - public static final int MASTERCOPROCESSORS_FIELD_NUMBER = 6; + // repeated .Coprocessor master_coprocessors = 6; + public static final int MASTER_COPROCESSORS_FIELD_NUMBER = 6; private java.util.List masterCoprocessors_; public java.util.List getMasterCoprocessorsList() { return masterCoprocessors_; @@ -4984,8 +4984,8 @@ public final class ClusterStatusProtos { return master_; } - // repeated .ServerName backupMasters = 8; - public static final int BACKUPMASTERS_FIELD_NUMBER = 8; + // repeated .ServerName backup_masters = 8; + public static final int BACKUP_MASTERS_FIELD_NUMBER = 8; private java.util.List backupMasters_; public java.util.List getBackupMastersList() { return backupMasters_; @@ -5005,8 +5005,8 @@ public final class ClusterStatusProtos { return backupMasters_.get(index); } - // optional bool balancerOn = 9; - public static final int BALANCERON_FIELD_NUMBER = 9; + // optional bool balancer_on = 9; + public static final int BALANCER_ON_FIELD_NUMBER = 9; private boolean balancerOn_; public boolean hasBalancerOn() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -5221,27 +5221,27 @@ public final class ClusterStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasHbaseVersion()) { - hash = (37 * hash) + HBASEVERSION_FIELD_NUMBER; + hash = (37 * hash) + HBASE_VERSION_FIELD_NUMBER; hash = (53 * hash) + getHbaseVersion().hashCode(); } if (getLiveServersCount() > 0) { - hash = (37 * hash) + LIVESERVERS_FIELD_NUMBER; + hash = (37 * hash) + LIVE_SERVERS_FIELD_NUMBER; hash = (53 * hash) + getLiveServersList().hashCode(); } if (getDeadServersCount() > 0) { - hash = (37 * hash) + DEADSERVERS_FIELD_NUMBER; + hash = (37 * hash) + DEAD_SERVERS_FIELD_NUMBER; hash = (53 * hash) + getDeadServersList().hashCode(); } if (getRegionsInTransitionCount() > 0) { - hash = (37 * hash) + REGIONSINTRANSITION_FIELD_NUMBER; + hash = (37 * hash) + REGIONS_IN_TRANSITION_FIELD_NUMBER; hash = (53 * hash) + getRegionsInTransitionList().hashCode(); } if (hasClusterId()) { - hash = (37 * hash) + CLUSTERID_FIELD_NUMBER; + hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER; hash = (53 * hash) + getClusterId().hashCode(); } if (getMasterCoprocessorsCount() > 0) { - hash = (37 * hash) + MASTERCOPROCESSORS_FIELD_NUMBER; + hash = (37 * hash) + MASTER_COPROCESSORS_FIELD_NUMBER; hash = (53 * hash) + getMasterCoprocessorsList().hashCode(); } if (hasMaster()) { @@ -5249,11 +5249,11 @@ public final class ClusterStatusProtos { hash = (53 * hash) + getMaster().hashCode(); } if (getBackupMastersCount() > 0) { - hash = (37 * hash) + BACKUPMASTERS_FIELD_NUMBER; + hash = (37 * hash) + BACKUP_MASTERS_FIELD_NUMBER; hash = (53 * hash) + getBackupMastersList().hashCode(); } if (hasBalancerOn()) { - hash = (37 * hash) + BALANCERON_FIELD_NUMBER; + hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getBalancerOn()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -5846,7 +5846,7 @@ public final class ClusterStatusProtos { private int bitField0_; - // optional .HBaseVersionFileContent hbaseVersion = 1; + // optional .HBaseVersionFileContent hbase_version = 1; private org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContent hbaseVersion_ = org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContent.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContent, org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContent.Builder, org.apache.hadoop.hbase.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder> hbaseVersionBuilder_; @@ -5936,7 +5936,7 @@ public final class ClusterStatusProtos { return hbaseVersionBuilder_; } - // repeated .LiveServerInfo liveServers = 2; + // repeated .LiveServerInfo live_servers = 2; private java.util.List liveServers_ = java.util.Collections.emptyList(); private void ensureLiveServersIsMutable() { @@ -6122,7 +6122,7 @@ public final class ClusterStatusProtos { return liveServersBuilder_; } - // repeated .ServerName deadServers = 3; + // repeated .ServerName dead_servers = 3; private java.util.List deadServers_ = java.util.Collections.emptyList(); private void ensureDeadServersIsMutable() { @@ -6308,7 +6308,7 @@ public final class ClusterStatusProtos { return deadServersBuilder_; } - // repeated .RegionInTransition regionsInTransition = 4; + // repeated .RegionInTransition regions_in_transition = 4; private java.util.List regionsInTransition_ = java.util.Collections.emptyList(); private void ensureRegionsInTransitionIsMutable() { @@ -6494,7 +6494,7 @@ public final class ClusterStatusProtos { return regionsInTransitionBuilder_; } - // optional .ClusterId clusterId = 5; + // optional .ClusterId cluster_id = 5; private org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterId clusterId_ = org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterId.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterId, org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterId.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterIdProtos.ClusterIdOrBuilder> clusterIdBuilder_; @@ -6584,7 +6584,7 @@ public final class ClusterStatusProtos { return clusterIdBuilder_; } - // repeated .Coprocessor masterCoprocessors = 6; + // repeated .Coprocessor master_coprocessors = 6; private java.util.List masterCoprocessors_ = java.util.Collections.emptyList(); private void ensureMasterCoprocessorsIsMutable() { @@ -6860,7 +6860,7 @@ public final class ClusterStatusProtos { return masterBuilder_; } - // repeated .ServerName backupMasters = 8; + // repeated .ServerName backup_masters = 8; private java.util.List backupMasters_ = java.util.Collections.emptyList(); private void ensureBackupMastersIsMutable() { @@ -7046,7 +7046,7 @@ public final class ClusterStatusProtos { return backupMastersBuilder_; } - // optional bool balancerOn = 9; + // optional bool balancer_on = 9; private boolean balancerOn_ ; public boolean hasBalancerOn() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -7118,46 +7118,48 @@ public final class ClusterStatusProtos { static { java.lang.String[] descriptorData = { "\n\023ClusterStatus.proto\032\013hbase.proto\032\017Clus" + - "terId.proto\032\010FS.proto\"\242\002\n\013RegionState\022\037\n" + - "\nregionInfo\030\001 \002(\0132\013.RegionInfo\022!\n\005state\030" + - "\002 \002(\0162\022.RegionState.State\022\r\n\005stamp\030\003 \001(\004" + - "\"\277\001\n\005State\022\013\n\007OFFLINE\020\000\022\020\n\014PENDING_OPEN\020" + - "\001\022\013\n\007OPENING\020\002\022\010\n\004OPEN\020\003\022\021\n\rPENDING_CLOS" + - "E\020\004\022\013\n\007CLOSING\020\005\022\n\n\006CLOSED\020\006\022\r\n\tSPLITTIN" + - "G\020\007\022\t\n\005SPLIT\020\010\022\017\n\013FAILED_OPEN\020\t\022\020\n\014FAILE" + - "D_CLOSE\020\n\022\013\n\007MERGING\020\013\022\n\n\006MERGED\020\014\"W\n\022Re" + - "gionInTransition\022\036\n\004spec\030\001 \002(\0132\020.RegionS", - "pecifier\022!\n\013regionState\030\002 \002(\0132\014.RegionSt" + - "ate\"\260\003\n\nRegionLoad\022)\n\017regionSpecifier\030\001 " + - "\002(\0132\020.RegionSpecifier\022\016\n\006stores\030\002 \001(\r\022\022\n" + - "\nstorefiles\030\003 \001(\r\022\037\n\027storeUncompressedSi" + - "zeMB\030\004 \001(\r\022\027\n\017storefileSizeMB\030\005 \001(\r\022\026\n\016m" + - "emstoreSizeMB\030\006 \001(\r\022\034\n\024storefileIndexSiz" + - "eMB\030\007 \001(\r\022\031\n\021readRequestsCount\030\010 \001(\004\022\032\n\022" + - "writeRequestsCount\030\t \001(\004\022\032\n\022totalCompact" + - "ingKVs\030\n \001(\004\022\033\n\023currentCompactedKVs\030\013 \001(" + - "\004\022\027\n\017rootIndexSizeKB\030\014 \001(\r\022\036\n\026totalStati", - "cIndexSizeKB\030\r \001(\r\022\036\n\026totalStaticBloomSi" + - "zeKB\030\016 \001(\r\022\032\n\022completeSequenceId\030\017 \001(\004\"\372" + - "\001\n\nServerLoad\022\030\n\020numberOfRequests\030\001 \001(\r\022" + - "\035\n\025totalNumberOfRequests\030\002 \001(\r\022\022\n\nusedHe" + - "apMB\030\003 \001(\r\022\021\n\tmaxHeapMB\030\004 \001(\r\022 \n\013regionL" + - "oads\030\005 \003(\0132\013.RegionLoad\022\"\n\014coprocessors\030" + - "\006 \003(\0132\014.Coprocessor\022\027\n\017reportStartTime\030\007" + - " \001(\004\022\025\n\rreportEndTime\030\010 \001(\004\022\026\n\016infoServe" + - "rPort\030\t \001(\r\"N\n\016LiveServerInfo\022\033\n\006server\030" + - "\001 \002(\0132\013.ServerName\022\037\n\nserverLoad\030\002 \002(\0132\013", - ".ServerLoad\"\327\002\n\rClusterStatus\022.\n\014hbaseVe" + - "rsion\030\001 \001(\0132\030.HBaseVersionFileContent\022$\n" + - "\013liveServers\030\002 \003(\0132\017.LiveServerInfo\022 \n\013d" + - "eadServers\030\003 \003(\0132\013.ServerName\0220\n\023regions" + - "InTransition\030\004 \003(\0132\023.RegionInTransition\022" + - "\035\n\tclusterId\030\005 \001(\0132\n.ClusterId\022(\n\022master" + - "Coprocessors\030\006 \003(\0132\014.Coprocessor\022\033\n\006mast" + - "er\030\007 \001(\0132\013.ServerName\022\"\n\rbackupMasters\030\010" + - " \003(\0132\013.ServerName\022\022\n\nbalancerOn\030\t \001(\010BF\n" + - "*org.apache.hadoop.hbase.protobuf.genera", - "tedB\023ClusterStatusProtosH\001\240\001\001" + "terId.proto\032\010FS.proto\"\243\002\n\013RegionState\022 \n" + + "\013region_info\030\001 \002(\0132\013.RegionInfo\022!\n\005state" + + "\030\002 \002(\0162\022.RegionState.State\022\r\n\005stamp\030\003 \001(" + + "\004\"\277\001\n\005State\022\013\n\007OFFLINE\020\000\022\020\n\014PENDING_OPEN" + + "\020\001\022\013\n\007OPENING\020\002\022\010\n\004OPEN\020\003\022\021\n\rPENDING_CLO" + + "SE\020\004\022\013\n\007CLOSING\020\005\022\n\n\006CLOSED\020\006\022\r\n\tSPLITTI" + + "NG\020\007\022\t\n\005SPLIT\020\010\022\017\n\013FAILED_OPEN\020\t\022\020\n\014FAIL" + + "ED_CLOSE\020\n\022\013\n\007MERGING\020\013\022\n\n\006MERGED\020\014\"X\n\022R" + + "egionInTransition\022\036\n\004spec\030\001 \002(\0132\020.Region", + "Specifier\022\"\n\014region_state\030\002 \002(\0132\014.Region" + + "State\"\320\003\n\nRegionLoad\022*\n\020region_specifier" + + "\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006stores\030\002 \001(\r" + + "\022\022\n\nstorefiles\030\003 \001(\r\022\"\n\032store_uncompress" + + "ed_size_MB\030\004 \001(\r\022\031\n\021storefile_size_MB\030\005 " + + "\001(\r\022\030\n\020memstore_size_MB\030\006 \001(\r\022\037\n\027storefi" + + "le_index_size_MB\030\007 \001(\r\022\033\n\023read_requests_" + + "count\030\010 \001(\004\022\034\n\024write_requests_count\030\t \001(" + + "\004\022\034\n\024total_compacting_KVs\030\n \001(\004\022\035\n\025curre" + + "nt_compacted_KVs\030\013 \001(\004\022\032\n\022root_index_siz", + "e_KB\030\014 \001(\r\022\"\n\032total_static_index_size_KB" + + "\030\r \001(\r\022\"\n\032total_static_bloom_size_KB\030\016 \001" + + "(\r\022\034\n\024complete_sequence_id\030\017 \001(\004\"\212\002\n\nSer" + + "verLoad\022\032\n\022number_of_requests\030\001 \001(\r\022 \n\030t" + + "otal_number_of_requests\030\002 \001(\r\022\024\n\014used_he" + + "ap_MB\030\003 \001(\r\022\023\n\013max_heap_MB\030\004 \001(\r\022!\n\014regi" + + "on_loads\030\005 \003(\0132\013.RegionLoad\022\"\n\014coprocess" + + "ors\030\006 \003(\0132\014.Coprocessor\022\031\n\021report_start_" + + "time\030\007 \001(\004\022\027\n\017report_end_time\030\010 \001(\004\022\030\n\020i" + + "nfo_server_port\030\t \001(\r\"O\n\016LiveServerInfo\022", + "\033\n\006server\030\001 \002(\0132\013.ServerName\022 \n\013server_l" + + "oad\030\002 \002(\0132\013.ServerLoad\"\340\002\n\rClusterStatus" + + "\022/\n\rhbase_version\030\001 \001(\0132\030.HBaseVersionFi" + + "leContent\022%\n\014live_servers\030\002 \003(\0132\017.LiveSe" + + "rverInfo\022!\n\014dead_servers\030\003 \003(\0132\013.ServerN" + + "ame\0222\n\025regions_in_transition\030\004 \003(\0132\023.Reg" + + "ionInTransition\022\036\n\ncluster_id\030\005 \001(\0132\n.Cl" + + "usterId\022)\n\023master_coprocessors\030\006 \003(\0132\014.C" + + "oprocessor\022\033\n\006master\030\007 \001(\0132\013.ServerName\022" + + "#\n\016backup_masters\030\010 \003(\0132\013.ServerName\022\023\n\013", + "balancer_on\030\t \001(\010BF\n*org.apache.hadoop.h" + + "base.protobuf.generatedB\023ClusterStatusPr" + + "otosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ComparatorProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ComparatorProtos.java index 983b57e..b9eadbd 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ComparatorProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ComparatorProtos.java @@ -15,7 +15,7 @@ public final class ComparatorProtos { boolean hasName(); String getName(); - // optional bytes serializedComparator = 2; + // optional bytes serialized_comparator = 2; boolean hasSerializedComparator(); com.google.protobuf.ByteString getSerializedComparator(); } @@ -80,8 +80,8 @@ public final class ComparatorProtos { } } - // optional bytes serializedComparator = 2; - public static final int SERIALIZEDCOMPARATOR_FIELD_NUMBER = 2; + // optional bytes serialized_comparator = 2; + public static final int SERIALIZED_COMPARATOR_FIELD_NUMBER = 2; private com.google.protobuf.ByteString serializedComparator_; public boolean hasSerializedComparator() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -180,7 +180,7 @@ public final class ComparatorProtos { hash = (53 * hash) + getName().hashCode(); } if (hasSerializedComparator()) { - hash = (37 * hash) + SERIALIZEDCOMPARATOR_FIELD_NUMBER; + hash = (37 * hash) + SERIALIZED_COMPARATOR_FIELD_NUMBER; hash = (53 * hash) + getSerializedComparator().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -458,7 +458,7 @@ public final class ComparatorProtos { onChanged(); } - // optional bytes serializedComparator = 2; + // optional bytes serialized_comparator = 2; private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY; public boolean hasSerializedComparator() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1805,7 +1805,7 @@ public final class ComparatorProtos { org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder(); - // required .BitComparator.BitwiseOp bitwiseOp = 2; + // required .BitComparator.BitwiseOp bitwise_op = 2; boolean hasBitwiseOp(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp(); } @@ -1923,8 +1923,8 @@ public final class ComparatorProtos { return comparable_; } - // required .BitComparator.BitwiseOp bitwiseOp = 2; - public static final int BITWISEOP_FIELD_NUMBER = 2; + // required .BitComparator.BitwiseOp bitwise_op = 2; + public static final int BITWISE_OP_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_; public boolean hasBitwiseOp() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2027,7 +2027,7 @@ public final class ComparatorProtos { hash = (53 * hash) + getComparable().hashCode(); } if (hasBitwiseOp()) { - hash = (37 * hash) + BITWISEOP_FIELD_NUMBER; + hash = (37 * hash) + BITWISE_OP_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getBitwiseOp()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2382,7 +2382,7 @@ public final class ComparatorProtos { return comparableBuilder_; } - // required .BitComparator.BitwiseOp bitwiseOp = 2; + // required .BitComparator.BitwiseOp bitwise_op = 2; private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND; public boolean hasBitwiseOp() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2724,7 +2724,7 @@ public final class ComparatorProtos { boolean hasPattern(); String getPattern(); - // required int32 patternFlags = 2; + // required int32 pattern_flags = 2; boolean hasPatternFlags(); int getPatternFlags(); @@ -2793,8 +2793,8 @@ public final class ComparatorProtos { } } - // required int32 patternFlags = 2; - public static final int PATTERNFLAGS_FIELD_NUMBER = 2; + // required int32 pattern_flags = 2; + public static final int PATTERN_FLAGS_FIELD_NUMBER = 2; private int patternFlags_; public boolean hasPatternFlags() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2946,7 +2946,7 @@ public final class ComparatorProtos { hash = (53 * hash) + getPattern().hashCode(); } if (hasPatternFlags()) { - hash = (37 * hash) + PATTERNFLAGS_FIELD_NUMBER; + hash = (37 * hash) + PATTERN_FLAGS_FIELD_NUMBER; hash = (53 * hash) + getPatternFlags(); } if (hasCharset()) { @@ -3250,7 +3250,7 @@ public final class ComparatorProtos { onChanged(); } - // required int32 patternFlags = 2; + // required int32 pattern_flags = 2; private int patternFlags_ ; public boolean hasPatternFlags() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3783,21 +3783,22 @@ public final class ComparatorProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\020Comparator.proto\"8\n\nComparator\022\014\n\004name" + - "\030\001 \002(\t\022\034\n\024serializedComparator\030\002 \001(\014\"$\n\023" + - "ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020Bi" + - "naryComparator\022(\n\ncomparable\030\001 \002(\0132\024.Byt" + - "eArrayComparable\"B\n\026BinaryPrefixComparat" + - "or\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompar" + - "able\"\215\001\n\rBitComparator\022(\n\ncomparable\030\001 \002" + - "(\0132\024.ByteArrayComparable\022+\n\tbitwiseOp\030\002 " + - "\002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwise" + - "Op\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullComp", - "arator\"O\n\025RegexStringComparator\022\017\n\007patte" + - "rn\030\001 \002(\t\022\024\n\014patternFlags\030\002 \002(\005\022\017\n\007charse" + - "t\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006substr" + - "\030\001 \002(\tBF\n*org.apache.hadoop.hbase.protob" + - "uf.generatedB\020ComparatorProtosH\001\210\001\001\240\001\001" + "\n\020Comparator.proto\"9\n\nComparator\022\014\n\004name" + + "\030\001 \002(\t\022\035\n\025serialized_comparator\030\002 \001(\014\"$\n" + + "\023ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020B" + + "inaryComparator\022(\n\ncomparable\030\001 \002(\0132\024.By" + + "teArrayComparable\"B\n\026BinaryPrefixCompara" + + "tor\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompa" + + "rable\"\216\001\n\rBitComparator\022(\n\ncomparable\030\001 " + + "\002(\0132\024.ByteArrayComparable\022,\n\nbitwise_op\030" + + "\002 \002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwi" + + "seOp\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullCo", + "mparator\"P\n\025RegexStringComparator\022\017\n\007pat" + + "tern\030\001 \002(\t\022\025\n\rpattern_flags\030\002 \002(\005\022\017\n\007cha" + + "rset\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006sub" + + "str\030\001 \002(\tBF\n*org.apache.hadoop.hbase.pro" + + "tobuf.generatedB\020ComparatorProtosH\001\210\001\001\240\001" + + "\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ErrorHandlingProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ErrorHandlingProtos.java index 067321f..c1b1660 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ErrorHandlingProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ErrorHandlingProtos.java @@ -11,19 +11,19 @@ public final class ErrorHandlingProtos { public interface StackTraceElementMessageOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional string declaringClass = 1; + // optional string declaring_class = 1; boolean hasDeclaringClass(); String getDeclaringClass(); - // optional string methodName = 2; + // optional string method_name = 2; boolean hasMethodName(); String getMethodName(); - // optional string fileName = 3; + // optional string file_name = 3; boolean hasFileName(); String getFileName(); - // optional int32 lineNumber = 4; + // optional int32 line_number = 4; boolean hasLineNumber(); int getLineNumber(); } @@ -56,8 +56,8 @@ public final class ErrorHandlingProtos { } private int bitField0_; - // optional string declaringClass = 1; - public static final int DECLARINGCLASS_FIELD_NUMBER = 1; + // optional string declaring_class = 1; + public static final int DECLARING_CLASS_FIELD_NUMBER = 1; private java.lang.Object declaringClass_; public boolean hasDeclaringClass() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -88,8 +88,8 @@ public final class ErrorHandlingProtos { } } - // optional string methodName = 2; - public static final int METHODNAME_FIELD_NUMBER = 2; + // optional string method_name = 2; + public static final int METHOD_NAME_FIELD_NUMBER = 2; private java.lang.Object methodName_; public boolean hasMethodName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -120,8 +120,8 @@ public final class ErrorHandlingProtos { } } - // optional string fileName = 3; - public static final int FILENAME_FIELD_NUMBER = 3; + // optional string file_name = 3; + public static final int FILE_NAME_FIELD_NUMBER = 3; private java.lang.Object fileName_; public boolean hasFileName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -152,8 +152,8 @@ public final class ErrorHandlingProtos { } } - // optional int32 lineNumber = 4; - public static final int LINENUMBER_FIELD_NUMBER = 4; + // optional int32 line_number = 4; + public static final int LINE_NUMBER_FIELD_NUMBER = 4; private int lineNumber_; public boolean hasLineNumber() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -270,19 +270,19 @@ public final class ErrorHandlingProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasDeclaringClass()) { - hash = (37 * hash) + DECLARINGCLASS_FIELD_NUMBER; + hash = (37 * hash) + DECLARING_CLASS_FIELD_NUMBER; hash = (53 * hash) + getDeclaringClass().hashCode(); } if (hasMethodName()) { - hash = (37 * hash) + METHODNAME_FIELD_NUMBER; + hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER; hash = (53 * hash) + getMethodName().hashCode(); } if (hasFileName()) { - hash = (37 * hash) + FILENAME_FIELD_NUMBER; + hash = (37 * hash) + FILE_NAME_FIELD_NUMBER; hash = (53 * hash) + getFileName().hashCode(); } if (hasLineNumber()) { - hash = (37 * hash) + LINENUMBER_FIELD_NUMBER; + hash = (37 * hash) + LINE_NUMBER_FIELD_NUMBER; hash = (53 * hash) + getLineNumber(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -548,7 +548,7 @@ public final class ErrorHandlingProtos { private int bitField0_; - // optional string declaringClass = 1; + // optional string declaring_class = 1; private java.lang.Object declaringClass_ = ""; public boolean hasDeclaringClass() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -584,7 +584,7 @@ public final class ErrorHandlingProtos { onChanged(); } - // optional string methodName = 2; + // optional string method_name = 2; private java.lang.Object methodName_ = ""; public boolean hasMethodName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -620,7 +620,7 @@ public final class ErrorHandlingProtos { onChanged(); } - // optional string fileName = 3; + // optional string file_name = 3; private java.lang.Object fileName_ = ""; public boolean hasFileName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -656,7 +656,7 @@ public final class ErrorHandlingProtos { onChanged(); } - // optional int32 lineNumber = 4; + // optional int32 line_number = 4; private int lineNumber_ ; public boolean hasLineNumber() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -691,7 +691,7 @@ public final class ErrorHandlingProtos { public interface GenericExceptionMessageOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional string className = 1; + // optional string class_name = 1; boolean hasClassName(); String getClassName(); @@ -699,7 +699,7 @@ public final class ErrorHandlingProtos { boolean hasMessage(); String getMessage(); - // optional bytes errorInfo = 3; + // optional bytes error_info = 3; boolean hasErrorInfo(); com.google.protobuf.ByteString getErrorInfo(); @@ -742,8 +742,8 @@ public final class ErrorHandlingProtos { } private int bitField0_; - // optional string className = 1; - public static final int CLASSNAME_FIELD_NUMBER = 1; + // optional string class_name = 1; + public static final int CLASS_NAME_FIELD_NUMBER = 1; private java.lang.Object className_; public boolean hasClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -806,8 +806,8 @@ public final class ErrorHandlingProtos { } } - // optional bytes errorInfo = 3; - public static final int ERRORINFO_FIELD_NUMBER = 3; + // optional bytes error_info = 3; + public static final int ERROR_INFO_FIELD_NUMBER = 3; private com.google.protobuf.ByteString errorInfo_; public boolean hasErrorInfo() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -942,7 +942,7 @@ public final class ErrorHandlingProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasClassName()) { - hash = (37 * hash) + CLASSNAME_FIELD_NUMBER; + hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getClassName().hashCode(); } if (hasMessage()) { @@ -950,7 +950,7 @@ public final class ErrorHandlingProtos { hash = (53 * hash) + getMessage().hashCode(); } if (hasErrorInfo()) { - hash = (37 * hash) + ERRORINFO_FIELD_NUMBER; + hash = (37 * hash) + ERROR_INFO_FIELD_NUMBER; hash = (53 * hash) + getErrorInfo().hashCode(); } if (getTraceCount() > 0) { @@ -1254,7 +1254,7 @@ public final class ErrorHandlingProtos { private int bitField0_; - // optional string className = 1; + // optional string class_name = 1; private java.lang.Object className_ = ""; public boolean hasClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1326,7 +1326,7 @@ public final class ErrorHandlingProtos { onChanged(); } - // optional bytes errorInfo = 3; + // optional bytes error_info = 3; private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY; public boolean hasErrorInfo() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1554,7 +1554,7 @@ public final class ErrorHandlingProtos { boolean hasSource(); String getSource(); - // optional .GenericExceptionMessage genericException = 2; + // optional .GenericExceptionMessage generic_exception = 2; boolean hasGenericException(); org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException(); org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder(); @@ -1620,8 +1620,8 @@ public final class ErrorHandlingProtos { } } - // optional .GenericExceptionMessage genericException = 2; - public static final int GENERICEXCEPTION_FIELD_NUMBER = 2; + // optional .GenericExceptionMessage generic_exception = 2; + public static final int GENERIC_EXCEPTION_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_; public boolean hasGenericException() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1719,7 +1719,7 @@ public final class ErrorHandlingProtos { hash = (53 * hash) + getSource().hashCode(); } if (hasGenericException()) { - hash = (37 * hash) + GENERICEXCEPTION_FIELD_NUMBER; + hash = (37 * hash) + GENERIC_EXCEPTION_FIELD_NUMBER; hash = (53 * hash) + getGenericException().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2006,7 +2006,7 @@ public final class ErrorHandlingProtos { onChanged(); } - // optional .GenericExceptionMessage genericException = 2; + // optional .GenericExceptionMessage generic_exception = 2; private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_; @@ -2131,17 +2131,17 @@ public final class ErrorHandlingProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\023ErrorHandling.proto\"l\n\030StackTraceEleme" + - "ntMessage\022\026\n\016declaringClass\030\001 \001(\t\022\022\n\nmet" + - "hodName\030\002 \001(\t\022\020\n\010fileName\030\003 \001(\t\022\022\n\nlineN" + - "umber\030\004 \001(\005\"z\n\027GenericExceptionMessage\022\021" + - "\n\tclassName\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022\021\n\ter" + - "rorInfo\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.StackTrac" + - "eElementMessage\"]\n\027ForeignExceptionMessa" + - "ge\022\016\n\006source\030\001 \001(\t\0222\n\020genericException\030\002" + - " \001(\0132\030.GenericExceptionMessageBF\n*org.ap" + - "ache.hadoop.hbase.protobuf.generatedB\023Er", - "rorHandlingProtosH\001\240\001\001" + "\n\023ErrorHandling.proto\"p\n\030StackTraceEleme" + + "ntMessage\022\027\n\017declaring_class\030\001 \001(\t\022\023\n\013me" + + "thod_name\030\002 \001(\t\022\021\n\tfile_name\030\003 \001(\t\022\023\n\013li" + + "ne_number\030\004 \001(\005\"|\n\027GenericExceptionMessa" + + "ge\022\022\n\nclass_name\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022" + + "\022\n\nerror_info\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.Sta" + + "ckTraceElementMessage\"^\n\027ForeignExceptio" + + "nMessage\022\016\n\006source\030\001 \001(\t\0223\n\021generic_exce" + + "ption\030\002 \001(\0132\030.GenericExceptionMessageBF\n" + + "*org.apache.hadoop.hbase.protobuf.genera", + "tedB\023ErrorHandlingProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java index fba358b..40bbab6 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/FilterProtos.java @@ -15,7 +15,7 @@ public final class FilterProtos { boolean hasName(); String getName(); - // optional bytes serializedFilter = 2; + // optional bytes serialized_filter = 2; boolean hasSerializedFilter(); com.google.protobuf.ByteString getSerializedFilter(); } @@ -80,8 +80,8 @@ public final class FilterProtos { } } - // optional bytes serializedFilter = 2; - public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2; + // optional bytes serialized_filter = 2; + public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2; private com.google.protobuf.ByteString serializedFilter_; public boolean hasSerializedFilter() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -180,7 +180,7 @@ public final class FilterProtos { hash = (53 * hash) + getName().hashCode(); } if (hasSerializedFilter()) { - hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER; + hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER; hash = (53 * hash) + getSerializedFilter().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -458,7 +458,7 @@ public final class FilterProtos { onChanged(); } - // optional bytes serializedFilter = 2; + // optional bytes serialized_filter = 2; private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY; public boolean hasSerializedFilter() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -883,7 +883,7 @@ public final class FilterProtos { boolean hasOffset(); int getOffset(); - // optional bytes columnOffset = 3; + // optional bytes column_offset = 3; boolean hasColumnOffset(); com.google.protobuf.ByteString getColumnOffset(); } @@ -936,8 +936,8 @@ public final class FilterProtos { return offset_; } - // optional bytes columnOffset = 3; - public static final int COLUMNOFFSET_FIELD_NUMBER = 3; + // optional bytes column_offset = 3; + public static final int COLUMN_OFFSET_FIELD_NUMBER = 3; private com.google.protobuf.ByteString columnOffset_; public boolean hasColumnOffset() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1053,7 +1053,7 @@ public final class FilterProtos { hash = (53 * hash) + getOffset(); } if (hasColumnOffset()) { - hash = (37 * hash) + COLUMNOFFSET_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER; hash = (53 * hash) + getColumnOffset().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1351,7 +1351,7 @@ public final class FilterProtos { return this; } - // optional bytes columnOffset = 3; + // optional bytes column_offset = 3; private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY; public boolean hasColumnOffset() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1771,19 +1771,19 @@ public final class FilterProtos { public interface ColumnRangeFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bytes minColumn = 1; + // optional bytes min_column = 1; boolean hasMinColumn(); com.google.protobuf.ByteString getMinColumn(); - // optional bool minColumnInclusive = 2; + // optional bool min_column_inclusive = 2; boolean hasMinColumnInclusive(); boolean getMinColumnInclusive(); - // optional bytes maxColumn = 3; + // optional bytes max_column = 3; boolean hasMaxColumn(); com.google.protobuf.ByteString getMaxColumn(); - // optional bool maxColumnInclusive = 4; + // optional bool max_column_inclusive = 4; boolean hasMaxColumnInclusive(); boolean getMaxColumnInclusive(); } @@ -1816,8 +1816,8 @@ public final class FilterProtos { } private int bitField0_; - // optional bytes minColumn = 1; - public static final int MINCOLUMN_FIELD_NUMBER = 1; + // optional bytes min_column = 1; + public static final int MIN_COLUMN_FIELD_NUMBER = 1; private com.google.protobuf.ByteString minColumn_; public boolean hasMinColumn() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1826,8 +1826,8 @@ public final class FilterProtos { return minColumn_; } - // optional bool minColumnInclusive = 2; - public static final int MINCOLUMNINCLUSIVE_FIELD_NUMBER = 2; + // optional bool min_column_inclusive = 2; + public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2; private boolean minColumnInclusive_; public boolean hasMinColumnInclusive() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1836,8 +1836,8 @@ public final class FilterProtos { return minColumnInclusive_; } - // optional bytes maxColumn = 3; - public static final int MAXCOLUMN_FIELD_NUMBER = 3; + // optional bytes max_column = 3; + public static final int MAX_COLUMN_FIELD_NUMBER = 3; private com.google.protobuf.ByteString maxColumn_; public boolean hasMaxColumn() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1846,8 +1846,8 @@ public final class FilterProtos { return maxColumn_; } - // optional bool maxColumnInclusive = 4; - public static final int MAXCOLUMNINCLUSIVE_FIELD_NUMBER = 4; + // optional bool max_column_inclusive = 4; + public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4; private boolean maxColumnInclusive_; public boolean hasMaxColumnInclusive() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1964,19 +1964,19 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasMinColumn()) { - hash = (37 * hash) + MINCOLUMN_FIELD_NUMBER; + hash = (37 * hash) + MIN_COLUMN_FIELD_NUMBER; hash = (53 * hash) + getMinColumn().hashCode(); } if (hasMinColumnInclusive()) { - hash = (37 * hash) + MINCOLUMNINCLUSIVE_FIELD_NUMBER; + hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getMinColumnInclusive()); } if (hasMaxColumn()) { - hash = (37 * hash) + MAXCOLUMN_FIELD_NUMBER; + hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER; hash = (53 * hash) + getMaxColumn().hashCode(); } if (hasMaxColumnInclusive()) { - hash = (37 * hash) + MAXCOLUMNINCLUSIVE_FIELD_NUMBER; + hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getMaxColumnInclusive()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2242,7 +2242,7 @@ public final class FilterProtos { private int bitField0_; - // optional bytes minColumn = 1; + // optional bytes min_column = 1; private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY; public boolean hasMinColumn() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2266,7 +2266,7 @@ public final class FilterProtos { return this; } - // optional bool minColumnInclusive = 2; + // optional bool min_column_inclusive = 2; private boolean minColumnInclusive_ ; public boolean hasMinColumnInclusive() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2287,7 +2287,7 @@ public final class FilterProtos { return this; } - // optional bytes maxColumn = 3; + // optional bytes max_column = 3; private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY; public boolean hasMaxColumn() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2311,7 +2311,7 @@ public final class FilterProtos { return this; } - // optional bool maxColumnInclusive = 4; + // optional bool max_column_inclusive = 4; private boolean maxColumnInclusive_ ; public boolean hasMaxColumnInclusive() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2346,7 +2346,7 @@ public final class FilterProtos { public interface CompareFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .CompareType compareOp = 1; + // required .CompareType compare_op = 1; boolean hasCompareOp(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp(); @@ -2384,8 +2384,8 @@ public final class FilterProtos { } private int bitField0_; - // required .CompareType compareOp = 1; - public static final int COMPAREOP_FIELD_NUMBER = 1; + // required .CompareType compare_op = 1; + public static final int COMPARE_OP_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_; public boolean hasCompareOp() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2499,7 +2499,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCompareOp()) { - hash = (37 * hash) + COMPAREOP_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getCompareOp()); } if (hasComparator()) { @@ -2770,7 +2770,7 @@ public final class FilterProtos { private int bitField0_; - // required .CompareType compareOp = 1; + // required .CompareType compare_op = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS; public boolean hasCompareOp() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2898,20 +2898,20 @@ public final class FilterProtos { public interface DependentColumnFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; boolean hasCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); - // optional bytes columnFamily = 2; + // optional bytes column_family = 2; boolean hasColumnFamily(); com.google.protobuf.ByteString getColumnFamily(); - // optional bytes columnQualifier = 3; + // optional bytes column_qualifier = 3; boolean hasColumnQualifier(); com.google.protobuf.ByteString getColumnQualifier(); - // optional bool dropDependentColumn = 4; + // optional bool drop_dependent_column = 4; boolean hasDropDependentColumn(); boolean getDropDependentColumn(); } @@ -2944,8 +2944,8 @@ public final class FilterProtos { } private int bitField0_; - // required .CompareFilter compareFilter = 1; - public static final int COMPAREFILTER_FIELD_NUMBER = 1; + // required .CompareFilter compare_filter = 1; + public static final int COMPARE_FILTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; public boolean hasCompareFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2957,8 +2957,8 @@ public final class FilterProtos { return compareFilter_; } - // optional bytes columnFamily = 2; - public static final int COLUMNFAMILY_FIELD_NUMBER = 2; + // optional bytes column_family = 2; + public static final int COLUMN_FAMILY_FIELD_NUMBER = 2; private com.google.protobuf.ByteString columnFamily_; public boolean hasColumnFamily() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2967,8 +2967,8 @@ public final class FilterProtos { return columnFamily_; } - // optional bytes columnQualifier = 3; - public static final int COLUMNQUALIFIER_FIELD_NUMBER = 3; + // optional bytes column_qualifier = 3; + public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 3; private com.google.protobuf.ByteString columnQualifier_; public boolean hasColumnQualifier() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2977,8 +2977,8 @@ public final class FilterProtos { return columnQualifier_; } - // optional bool dropDependentColumn = 4; - public static final int DROPDEPENDENTCOLUMN_FIELD_NUMBER = 4; + // optional bool drop_dependent_column = 4; + public static final int DROP_DEPENDENT_COLUMN_FIELD_NUMBER = 4; private boolean dropDependentColumn_; public boolean hasDropDependentColumn() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -3103,19 +3103,19 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCompareFilter()) { - hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER; hash = (53 * hash) + getCompareFilter().hashCode(); } if (hasColumnFamily()) { - hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER; hash = (53 * hash) + getColumnFamily().hashCode(); } if (hasColumnQualifier()) { - hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER; hash = (53 * hash) + getColumnQualifier().hashCode(); } if (hasDropDependentColumn()) { - hash = (37 * hash) + DROPDEPENDENTCOLUMN_FIELD_NUMBER; + hash = (37 * hash) + DROP_DEPENDENT_COLUMN_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getDropDependentColumn()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3402,7 +3402,7 @@ public final class FilterProtos { private int bitField0_; - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; @@ -3492,7 +3492,7 @@ public final class FilterProtos { return compareFilterBuilder_; } - // optional bytes columnFamily = 2; + // optional bytes column_family = 2; private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY; public boolean hasColumnFamily() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3516,7 +3516,7 @@ public final class FilterProtos { return this; } - // optional bytes columnQualifier = 3; + // optional bytes column_qualifier = 3; private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY; public boolean hasColumnQualifier() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -3540,7 +3540,7 @@ public final class FilterProtos { return this; } - // optional bool dropDependentColumn = 4; + // optional bool drop_dependent_column = 4; private boolean dropDependentColumn_ ; public boolean hasDropDependentColumn() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -3575,7 +3575,7 @@ public final class FilterProtos { public interface FamilyFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; boolean hasCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); @@ -3609,8 +3609,8 @@ public final class FilterProtos { } private int bitField0_; - // required .CompareFilter compareFilter = 1; - public static final int COMPAREFILTER_FIELD_NUMBER = 1; + // required .CompareFilter compare_filter = 1; + public static final int COMPARE_FILTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; public boolean hasCompareFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3699,7 +3699,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCompareFilter()) { - hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER; hash = (53 * hash) + getCompareFilter().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3944,7 +3944,7 @@ public final class FilterProtos { private int bitField0_; - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; @@ -5982,7 +5982,7 @@ public final class FilterProtos { public interface FuzzyRowFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .BytesBytesPair fuzzyKeysData = 1; + // repeated .BytesBytesPair fuzzy_keys_data = 1; java.util.List getFuzzyKeysDataList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index); @@ -6020,8 +6020,8 @@ public final class FilterProtos { return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable; } - // repeated .BytesBytesPair fuzzyKeysData = 1; - public static final int FUZZYKEYSDATA_FIELD_NUMBER = 1; + // repeated .BytesBytesPair fuzzy_keys_data = 1; + public static final int FUZZY_KEYS_DATA_FIELD_NUMBER = 1; private java.util.List fuzzyKeysData_; public java.util.List getFuzzyKeysDataList() { return fuzzyKeysData_; @@ -6113,7 +6113,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getFuzzyKeysDataCount() > 0) { - hash = (37 * hash) + FUZZYKEYSDATA_FIELD_NUMBER; + hash = (37 * hash) + FUZZY_KEYS_DATA_FIELD_NUMBER; hash = (53 * hash) + getFuzzyKeysDataList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -6375,7 +6375,7 @@ public final class FilterProtos { private int bitField0_; - // repeated .BytesBytesPair fuzzyKeysData = 1; + // repeated .BytesBytesPair fuzzy_keys_data = 1; private java.util.List fuzzyKeysData_ = java.util.Collections.emptyList(); private void ensureFuzzyKeysDataIsMutable() { @@ -6575,7 +6575,7 @@ public final class FilterProtos { public interface InclusiveStopFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bytes stopRowKey = 1; + // optional bytes stop_row_key = 1; boolean hasStopRowKey(); com.google.protobuf.ByteString getStopRowKey(); } @@ -6608,8 +6608,8 @@ public final class FilterProtos { } private int bitField0_; - // optional bytes stopRowKey = 1; - public static final int STOPROWKEY_FIELD_NUMBER = 1; + // optional bytes stop_row_key = 1; + public static final int STOP_ROW_KEY_FIELD_NUMBER = 1; private com.google.protobuf.ByteString stopRowKey_; public boolean hasStopRowKey() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6687,7 +6687,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasStopRowKey()) { - hash = (37 * hash) + STOPROWKEY_FIELD_NUMBER; + hash = (37 * hash) + STOP_ROW_KEY_FIELD_NUMBER; hash = (53 * hash) + getStopRowKey().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -6911,7 +6911,7 @@ public final class FilterProtos { private int bitField0_; - // optional bytes stopRowKey = 1; + // optional bytes stop_row_key = 1; private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY; public boolean hasStopRowKey() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6949,7 +6949,7 @@ public final class FilterProtos { public interface KeyOnlyFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bool lenAsVal = 1; + // required bool len_as_val = 1; boolean hasLenAsVal(); boolean getLenAsVal(); } @@ -6982,8 +6982,8 @@ public final class FilterProtos { } private int bitField0_; - // required bool lenAsVal = 1; - public static final int LENASVAL_FIELD_NUMBER = 1; + // required bool len_as_val = 1; + public static final int LEN_AS_VAL_FIELD_NUMBER = 1; private boolean lenAsVal_; public boolean hasLenAsVal() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -7065,7 +7065,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLenAsVal()) { - hash = (37 * hash) + LENASVAL_FIELD_NUMBER; + hash = (37 * hash) + LEN_AS_VAL_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getLenAsVal()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -7293,7 +7293,7 @@ public final class FilterProtos { private int bitField0_; - // required bool lenAsVal = 1; + // required bool len_as_val = 1; private boolean lenAsVal_ ; public boolean hasLenAsVal() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -7328,7 +7328,7 @@ public final class FilterProtos { public interface MultipleColumnPrefixFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated bytes sortedPrefixes = 1; + // repeated bytes sorted_prefixes = 1; java.util.List getSortedPrefixesList(); int getSortedPrefixesCount(); com.google.protobuf.ByteString getSortedPrefixes(int index); @@ -7361,8 +7361,8 @@ public final class FilterProtos { return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable; } - // repeated bytes sortedPrefixes = 1; - public static final int SORTEDPREFIXES_FIELD_NUMBER = 1; + // repeated bytes sorted_prefixes = 1; + public static final int SORTED_PREFIXES_FIELD_NUMBER = 1; private java.util.List sortedPrefixes_; public java.util.List getSortedPrefixesList() { @@ -7446,7 +7446,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getSortedPrefixesCount() > 0) { - hash = (37 * hash) + SORTEDPREFIXES_FIELD_NUMBER; + hash = (37 * hash) + SORTED_PREFIXES_FIELD_NUMBER; hash = (53 * hash) + getSortedPrefixesList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -7676,7 +7676,7 @@ public final class FilterProtos { private int bitField0_; - // repeated bytes sortedPrefixes = 1; + // repeated bytes sorted_prefixes = 1; private java.util.List sortedPrefixes_ = java.util.Collections.emptyList();; private void ensureSortedPrefixesIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { @@ -7741,7 +7741,7 @@ public final class FilterProtos { public interface PageFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required int64 pageSize = 1; + // required int64 page_size = 1; boolean hasPageSize(); long getPageSize(); } @@ -7774,8 +7774,8 @@ public final class FilterProtos { } private int bitField0_; - // required int64 pageSize = 1; - public static final int PAGESIZE_FIELD_NUMBER = 1; + // required int64 page_size = 1; + public static final int PAGE_SIZE_FIELD_NUMBER = 1; private long pageSize_; public boolean hasPageSize() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -7857,7 +7857,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasPageSize()) { - hash = (37 * hash) + PAGESIZE_FIELD_NUMBER; + hash = (37 * hash) + PAGE_SIZE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getPageSize()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -8085,7 +8085,7 @@ public final class FilterProtos { private int bitField0_; - // required int64 pageSize = 1; + // required int64 page_size = 1; private long pageSize_ ; public boolean hasPageSize() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -8494,7 +8494,7 @@ public final class FilterProtos { public interface QualifierFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; boolean hasCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); @@ -8528,8 +8528,8 @@ public final class FilterProtos { } private int bitField0_; - // required .CompareFilter compareFilter = 1; - public static final int COMPAREFILTER_FIELD_NUMBER = 1; + // required .CompareFilter compare_filter = 1; + public static final int COMPARE_FILTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; public boolean hasCompareFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -8618,7 +8618,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCompareFilter()) { - hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER; hash = (53 * hash) + getCompareFilter().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -8863,7 +8863,7 @@ public final class FilterProtos { private int bitField0_; - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; @@ -9346,7 +9346,7 @@ public final class FilterProtos { public interface RowFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; boolean hasCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); @@ -9380,8 +9380,8 @@ public final class FilterProtos { } private int bitField0_; - // required .CompareFilter compareFilter = 1; - public static final int COMPAREFILTER_FIELD_NUMBER = 1; + // required .CompareFilter compare_filter = 1; + public static final int COMPARE_FILTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; public boolean hasCompareFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9470,7 +9470,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCompareFilter()) { - hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER; hash = (53 * hash) + getCompareFilter().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -9715,7 +9715,7 @@ public final class FilterProtos { private int bitField0_; - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; @@ -9819,7 +9819,7 @@ public final class FilterProtos { public interface SingleColumnValueExcludeFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .SingleColumnValueFilter singleColumnValueFilter = 1; + // required .SingleColumnValueFilter single_column_value_filter = 1; boolean hasSingleColumnValueFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder(); @@ -9853,8 +9853,8 @@ public final class FilterProtos { } private int bitField0_; - // required .SingleColumnValueFilter singleColumnValueFilter = 1; - public static final int SINGLECOLUMNVALUEFILTER_FIELD_NUMBER = 1; + // required .SingleColumnValueFilter single_column_value_filter = 1; + public static final int SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_; public boolean hasSingleColumnValueFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9943,7 +9943,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasSingleColumnValueFilter()) { - hash = (37 * hash) + SINGLECOLUMNVALUEFILTER_FIELD_NUMBER; + hash = (37 * hash) + SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER; hash = (53 * hash) + getSingleColumnValueFilter().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -10188,7 +10188,7 @@ public final class FilterProtos { private int bitField0_; - // required .SingleColumnValueFilter singleColumnValueFilter = 1; + // required .SingleColumnValueFilter single_column_value_filter = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_; @@ -10292,15 +10292,15 @@ public final class FilterProtos { public interface SingleColumnValueFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bytes columnFamily = 1; + // optional bytes column_family = 1; boolean hasColumnFamily(); com.google.protobuf.ByteString getColumnFamily(); - // optional bytes columnQualifier = 2; + // optional bytes column_qualifier = 2; boolean hasColumnQualifier(); com.google.protobuf.ByteString getColumnQualifier(); - // required .CompareType compareOp = 3; + // required .CompareType compare_op = 3; boolean hasCompareOp(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp(); @@ -10309,11 +10309,11 @@ public final class FilterProtos { org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder(); - // optional bool filterIfMissing = 5; + // optional bool filter_if_missing = 5; boolean hasFilterIfMissing(); boolean getFilterIfMissing(); - // optional bool latestVersionOnly = 6; + // optional bool latest_version_only = 6; boolean hasLatestVersionOnly(); boolean getLatestVersionOnly(); } @@ -10346,8 +10346,8 @@ public final class FilterProtos { } private int bitField0_; - // optional bytes columnFamily = 1; - public static final int COLUMNFAMILY_FIELD_NUMBER = 1; + // optional bytes column_family = 1; + public static final int COLUMN_FAMILY_FIELD_NUMBER = 1; private com.google.protobuf.ByteString columnFamily_; public boolean hasColumnFamily() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -10356,8 +10356,8 @@ public final class FilterProtos { return columnFamily_; } - // optional bytes columnQualifier = 2; - public static final int COLUMNQUALIFIER_FIELD_NUMBER = 2; + // optional bytes column_qualifier = 2; + public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 2; private com.google.protobuf.ByteString columnQualifier_; public boolean hasColumnQualifier() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -10366,8 +10366,8 @@ public final class FilterProtos { return columnQualifier_; } - // required .CompareType compareOp = 3; - public static final int COMPAREOP_FIELD_NUMBER = 3; + // required .CompareType compare_op = 3; + public static final int COMPARE_OP_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_; public boolean hasCompareOp() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -10389,8 +10389,8 @@ public final class FilterProtos { return comparator_; } - // optional bool filterIfMissing = 5; - public static final int FILTERIFMISSING_FIELD_NUMBER = 5; + // optional bool filter_if_missing = 5; + public static final int FILTER_IF_MISSING_FIELD_NUMBER = 5; private boolean filterIfMissing_; public boolean hasFilterIfMissing() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -10399,8 +10399,8 @@ public final class FilterProtos { return filterIfMissing_; } - // optional bool latestVersionOnly = 6; - public static final int LATESTVERSIONONLY_FIELD_NUMBER = 6; + // optional bool latest_version_only = 6; + public static final int LATEST_VERSION_ONLY_FIELD_NUMBER = 6; private boolean latestVersionOnly_; public boolean hasLatestVersionOnly() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -10555,15 +10555,15 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasColumnFamily()) { - hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER; hash = (53 * hash) + getColumnFamily().hashCode(); } if (hasColumnQualifier()) { - hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER; hash = (53 * hash) + getColumnQualifier().hashCode(); } if (hasCompareOp()) { - hash = (37 * hash) + COMPAREOP_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getCompareOp()); } if (hasComparator()) { @@ -10571,11 +10571,11 @@ public final class FilterProtos { hash = (53 * hash) + getComparator().hashCode(); } if (hasFilterIfMissing()) { - hash = (37 * hash) + FILTERIFMISSING_FIELD_NUMBER; + hash = (37 * hash) + FILTER_IF_MISSING_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getFilterIfMissing()); } if (hasLatestVersionOnly()) { - hash = (37 * hash) + LATESTVERSIONONLY_FIELD_NUMBER; + hash = (37 * hash) + LATEST_VERSION_ONLY_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getLatestVersionOnly()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -10900,7 +10900,7 @@ public final class FilterProtos { private int bitField0_; - // optional bytes columnFamily = 1; + // optional bytes column_family = 1; private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY; public boolean hasColumnFamily() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -10924,7 +10924,7 @@ public final class FilterProtos { return this; } - // optional bytes columnQualifier = 2; + // optional bytes column_qualifier = 2; private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY; public boolean hasColumnQualifier() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -10948,7 +10948,7 @@ public final class FilterProtos { return this; } - // required .CompareType compareOp = 3; + // required .CompareType compare_op = 3; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS; public boolean hasCompareOp() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -11062,7 +11062,7 @@ public final class FilterProtos { return comparatorBuilder_; } - // optional bool filterIfMissing = 5; + // optional bool filter_if_missing = 5; private boolean filterIfMissing_ ; public boolean hasFilterIfMissing() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -11083,7 +11083,7 @@ public final class FilterProtos { return this; } - // optional bool latestVersionOnly = 6; + // optional bool latest_version_only = 6; private boolean latestVersionOnly_ ; public boolean hasLatestVersionOnly() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -12017,7 +12017,7 @@ public final class FilterProtos { public interface ValueFilterOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; boolean hasCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); @@ -12051,8 +12051,8 @@ public final class FilterProtos { } private int bitField0_; - // required .CompareFilter compareFilter = 1; - public static final int COMPAREFILTER_FIELD_NUMBER = 1; + // required .CompareFilter compare_filter = 1; + public static final int COMPARE_FILTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; public boolean hasCompareFilter() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -12141,7 +12141,7 @@ public final class FilterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCompareFilter()) { - hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; + hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER; hash = (53 * hash) + getCompareFilter().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -12386,7 +12386,7 @@ public final class FilterProtos { private int bitField0_; - // required .CompareFilter compareFilter = 1; + // required .CompareFilter compare_filter = 1; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; @@ -13105,51 +13105,52 @@ public final class FilterProtos { static { java.lang.String[] descriptorData = { "\n\014Filter.proto\032\013hbase.proto\032\020Comparator." + - "proto\"0\n\006Filter\022\014\n\004name\030\001 \002(\t\022\030\n\020seriali" + - "zedFilter\030\002 \001(\014\"%\n\024ColumnCountGetFilter\022" + - "\r\n\005limit\030\001 \002(\005\"M\n\026ColumnPaginationFilter" + - "\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014colum" + - "nOffset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006p" + - "refix\030\001 \002(\014\"q\n\021ColumnRangeFilter\022\021\n\tminC" + - "olumn\030\001 \001(\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022" + - "\021\n\tmaxColumn\030\003 \001(\014\022\032\n\022maxColumnInclusive" + - "\030\004 \001(\010\"Q\n\rCompareFilter\022\037\n\tcompareOp\030\001 \002", - "(\0162\014.CompareType\022\037\n\ncomparator\030\002 \001(\0132\013.C" + - "omparator\"\212\001\n\025DependentColumnFilter\022%\n\rc" + - "ompareFilter\030\001 \002(\0132\016.CompareFilter\022\024\n\014co" + - "lumnFamily\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(" + - "\014\022\033\n\023dropDependentColumn\030\004 \001(\010\"5\n\014Family" + - "Filter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareF" + - "ilter\"\200\001\n\nFilterList\022&\n\010operator\030\001 \002(\0162\024" + - ".FilterList.Operator\022\030\n\007filters\030\002 \003(\0132\007." + - "Filter\"0\n\010Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n" + - "\rMUST_PASS_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006fil", - "ter\030\001 \002(\0132\007.Filter\"\024\n\022FirstKeyOnlyFilter" + - "\";\n%FirstKeyValueMatchingQualifiersFilte" + - "r\022\022\n\nqualifiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022" + - "&\n\rfuzzyKeysData\030\001 \003(\0132\017.BytesBytesPair\"" + - ")\n\023InclusiveStopFilter\022\022\n\nstopRowKey\030\001 \001" + - "(\014\"!\n\rKeyOnlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n" + - "\032MultipleColumnPrefixFilter\022\026\n\016sortedPre" + - "fixes\030\001 \003(\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 " + - "\002(\003\"\036\n\014PrefixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017Q" + - "ualifierFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.", - "CompareFilter\"!\n\017RandomRowFilter\022\016\n\006chan" + - "ce\030\001 \002(\002\"2\n\tRowFilter\022%\n\rcompareFilter\030\001" + - " \002(\0132\016.CompareFilter\"[\n\036SingleColumnValu" + - "eExcludeFilter\0229\n\027singleColumnValueFilte" + - "r\030\001 \002(\0132\030.SingleColumnValueFilter\"\276\001\n\027Si" + - "ngleColumnValueFilter\022\024\n\014columnFamily\030\001 " + - "\001(\014\022\027\n\017columnQualifier\030\002 \001(\014\022\037\n\tcompareO" + - "p\030\003 \002(\0162\014.CompareType\022\037\n\ncomparator\030\004 \002(" + - "\0132\013.Comparator\022\027\n\017filterIfMissing\030\005 \001(\010\022" + - "\031\n\021latestVersionOnly\030\006 \001(\010\"%\n\nSkipFilter", - "\022\027\n\006filter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsF" + - "ilter\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFi" + - "lter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareFil" + - "ter\"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132" + - "\007.FilterBB\n*org.apache.hadoop.hbase.prot" + - "obuf.generatedB\014FilterProtosH\001\210\001\001\240\001\001" + "proto\"1\n\006Filter\022\014\n\004name\030\001 \002(\t\022\031\n\021seriali" + + "zed_filter\030\002 \001(\014\"%\n\024ColumnCountGetFilter" + + "\022\r\n\005limit\030\001 \002(\005\"N\n\026ColumnPaginationFilte" + + "r\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\025\n\rcolu" + + "mn_offset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n" + + "\006prefix\030\001 \002(\014\"w\n\021ColumnRangeFilter\022\022\n\nmi" + + "n_column\030\001 \001(\014\022\034\n\024min_column_inclusive\030\002" + + " \001(\010\022\022\n\nmax_column\030\003 \001(\014\022\034\n\024max_column_i" + + "nclusive\030\004 \001(\010\"R\n\rCompareFilter\022 \n\ncompa", + "re_op\030\001 \002(\0162\014.CompareType\022\037\n\ncomparator\030" + + "\002 \001(\0132\013.Comparator\"\217\001\n\025DependentColumnFi" + + "lter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareFi" + + "lter\022\025\n\rcolumn_family\030\002 \001(\014\022\030\n\020column_qu" + + "alifier\030\003 \001(\014\022\035\n\025drop_dependent_column\030\004" + + " \001(\010\"6\n\014FamilyFilter\022&\n\016compare_filter\030\001" + + " \002(\0132\016.CompareFilter\"\200\001\n\nFilterList\022&\n\010o" + + "perator\030\001 \002(\0162\024.FilterList.Operator\022\030\n\007f" + + "ilters\030\002 \003(\0132\007.Filter\"0\n\010Operator\022\021\n\rMUS" + + "T_PASS_ALL\020\001\022\021\n\rMUST_PASS_ONE\020\002\"(\n\rFilte", + "rWrapper\022\027\n\006filter\030\001 \002(\0132\007.Filter\"\024\n\022Fir" + + "stKeyOnlyFilter\";\n%FirstKeyValueMatching" + + "QualifiersFilter\022\022\n\nqualifiers\030\001 \003(\014\":\n\016" + + "FuzzyRowFilter\022(\n\017fuzzy_keys_data\030\001 \003(\0132" + + "\017.BytesBytesPair\"+\n\023InclusiveStopFilter\022" + + "\024\n\014stop_row_key\030\001 \001(\014\"#\n\rKeyOnlyFilter\022\022" + + "\n\nlen_as_val\030\001 \002(\010\"5\n\032MultipleColumnPref" + + "ixFilter\022\027\n\017sorted_prefixes\030\001 \003(\014\"\037\n\nPag" + + "eFilter\022\021\n\tpage_size\030\001 \002(\003\"\036\n\014PrefixFilt" + + "er\022\016\n\006prefix\030\001 \001(\014\"9\n\017QualifierFilter\022&\n", + "\016compare_filter\030\001 \002(\0132\016.CompareFilter\"!\n" + + "\017RandomRowFilter\022\016\n\006chance\030\001 \002(\002\"3\n\tRowF" + + "ilter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareF" + + "ilter\"^\n\036SingleColumnValueExcludeFilter\022" + + "<\n\032single_column_value_filter\030\001 \002(\0132\030.Si" + + "ngleColumnValueFilter\"\305\001\n\027SingleColumnVa" + + "lueFilter\022\025\n\rcolumn_family\030\001 \001(\014\022\030\n\020colu" + + "mn_qualifier\030\002 \001(\014\022 \n\ncompare_op\030\003 \002(\0162\014" + + ".CompareType\022\037\n\ncomparator\030\004 \002(\0132\013.Compa" + + "rator\022\031\n\021filter_if_missing\030\005 \001(\010\022\033\n\023late", + "st_version_only\030\006 \001(\010\"%\n\nSkipFilter\022\027\n\006f" + + "ilter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsFilter" + + "\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"5\n\013ValueFilter\022" + + "&\n\016compare_filter\030\001 \002(\0132\016.CompareFilter\"" + + "+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132\007.Fi" + + "lterBB\n*org.apache.hadoop.hbase.protobuf" + + ".generatedB\014FilterProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java index 1ba0177..d993300 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java @@ -109,7 +109,7 @@ public final class HBaseProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( int index); - // repeated .ColumnFamilySchema columnFamilies = 3; + // repeated .ColumnFamilySchema column_families = 3; java.util.List getColumnFamiliesList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index); @@ -189,8 +189,8 @@ public final class HBaseProtos { return attributes_.get(index); } - // repeated .ColumnFamilySchema columnFamilies = 3; - public static final int COLUMNFAMILIES_FIELD_NUMBER = 3; + // repeated .ColumnFamilySchema column_families = 3; + public static final int COLUMN_FAMILIES_FIELD_NUMBER = 3; private java.util.List columnFamilies_; public java.util.List getColumnFamiliesList() { return columnFamilies_; @@ -356,7 +356,7 @@ public final class HBaseProtos { hash = (53 * hash) + getAttributesList().hashCode(); } if (getColumnFamiliesCount() > 0) { - hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER; hash = (53 * hash) + getColumnFamiliesList().hashCode(); } if (getConfigurationCount() > 0) { @@ -956,7 +956,7 @@ public final class HBaseProtos { return attributesBuilder_; } - // repeated .ColumnFamilySchema columnFamilies = 3; + // repeated .ColumnFamilySchema column_families = 3; private java.util.List columnFamilies_ = java.util.Collections.emptyList(); private void ensureColumnFamiliesIsMutable() { @@ -2306,19 +2306,19 @@ public final class HBaseProtos { public interface RegionInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required uint64 regionId = 1; + // required uint64 region_id = 1; boolean hasRegionId(); long getRegionId(); - // required bytes tableName = 2; + // required bytes table_name = 2; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // optional bytes startKey = 3; + // optional bytes start_key = 3; boolean hasStartKey(); com.google.protobuf.ByteString getStartKey(); - // optional bytes endKey = 4; + // optional bytes end_key = 4; boolean hasEndKey(); com.google.protobuf.ByteString getEndKey(); @@ -2363,8 +2363,8 @@ public final class HBaseProtos { } private int bitField0_; - // required uint64 regionId = 1; - public static final int REGIONID_FIELD_NUMBER = 1; + // required uint64 region_id = 1; + public static final int REGION_ID_FIELD_NUMBER = 1; private long regionId_; public boolean hasRegionId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2373,8 +2373,8 @@ public final class HBaseProtos { return regionId_; } - // required bytes tableName = 2; - public static final int TABLENAME_FIELD_NUMBER = 2; + // required bytes table_name = 2; + public static final int TABLE_NAME_FIELD_NUMBER = 2; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2383,8 +2383,8 @@ public final class HBaseProtos { return tableName_; } - // optional bytes startKey = 3; - public static final int STARTKEY_FIELD_NUMBER = 3; + // optional bytes start_key = 3; + public static final int START_KEY_FIELD_NUMBER = 3; private com.google.protobuf.ByteString startKey_; public boolean hasStartKey() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2393,8 +2393,8 @@ public final class HBaseProtos { return startKey_; } - // optional bytes endKey = 4; - public static final int ENDKEY_FIELD_NUMBER = 4; + // optional bytes end_key = 4; + public static final int END_KEY_FIELD_NUMBER = 4; private com.google.protobuf.ByteString endKey_; public boolean hasEndKey() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2588,19 +2588,19 @@ public final class HBaseProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionId()) { - hash = (37 * hash) + REGIONID_FIELD_NUMBER; + hash = (37 * hash) + REGION_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getRegionId()); } if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasStartKey()) { - hash = (37 * hash) + STARTKEY_FIELD_NUMBER; + hash = (37 * hash) + START_KEY_FIELD_NUMBER; hash = (53 * hash) + getStartKey().hashCode(); } if (hasEndKey()) { - hash = (37 * hash) + ENDKEY_FIELD_NUMBER; + hash = (37 * hash) + END_KEY_FIELD_NUMBER; hash = (53 * hash) + getEndKey().hashCode(); } if (hasOffline()) { @@ -2928,7 +2928,7 @@ public final class HBaseProtos { private int bitField0_; - // required uint64 regionId = 1; + // required uint64 region_id = 1; private long regionId_ ; public boolean hasRegionId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2949,7 +2949,7 @@ public final class HBaseProtos { return this; } - // required bytes tableName = 2; + // required bytes table_name = 2; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2973,7 +2973,7 @@ public final class HBaseProtos { return this; } - // optional bytes startKey = 3; + // optional bytes start_key = 3; private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY; public boolean hasStartKey() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2997,7 +2997,7 @@ public final class HBaseProtos { return this; } - // optional bytes endKey = 4; + // optional bytes end_key = 4; private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY; public boolean hasEndKey() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -3098,7 +3098,7 @@ public final class HBaseProtos { public interface FavoredNodesOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .ServerName favoredNode = 1; + // repeated .ServerName favored_node = 1; java.util.List getFavoredNodeList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index); @@ -3136,8 +3136,8 @@ public final class HBaseProtos { return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable; } - // repeated .ServerName favoredNode = 1; - public static final int FAVOREDNODE_FIELD_NUMBER = 1; + // repeated .ServerName favored_node = 1; + public static final int FAVORED_NODE_FIELD_NUMBER = 1; private java.util.List favoredNode_; public java.util.List getFavoredNodeList() { return favoredNode_; @@ -3229,7 +3229,7 @@ public final class HBaseProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getFavoredNodeCount() > 0) { - hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER; + hash = (37 * hash) + FAVORED_NODE_FIELD_NUMBER; hash = (53 * hash) + getFavoredNodeList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3491,7 +3491,7 @@ public final class HBaseProtos { private int bitField0_; - // repeated .ServerName favoredNode = 1; + // repeated .ServerName favored_node = 1; private java.util.List favoredNode_ = java.util.Collections.emptyList(); private void ensureFavoredNodeIsMutable() { @@ -4662,7 +4662,7 @@ public final class HBaseProtos { public interface ServerNameOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string hostName = 1; + // required string host_name = 1; boolean hasHostName(); String getHostName(); @@ -4670,7 +4670,7 @@ public final class HBaseProtos { boolean hasPort(); int getPort(); - // optional uint64 startCode = 3; + // optional uint64 start_code = 3; boolean hasStartCode(); long getStartCode(); } @@ -4703,8 +4703,8 @@ public final class HBaseProtos { } private int bitField0_; - // required string hostName = 1; - public static final int HOSTNAME_FIELD_NUMBER = 1; + // required string host_name = 1; + public static final int HOST_NAME_FIELD_NUMBER = 1; private java.lang.Object hostName_; public boolean hasHostName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -4745,8 +4745,8 @@ public final class HBaseProtos { return port_; } - // optional uint64 startCode = 3; - public static final int STARTCODE_FIELD_NUMBER = 3; + // optional uint64 start_code = 3; + public static final int START_CODE_FIELD_NUMBER = 3; private long startCode_; public boolean hasStartCode() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -4854,7 +4854,7 @@ public final class HBaseProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasHostName()) { - hash = (37 * hash) + HOSTNAME_FIELD_NUMBER; + hash = (37 * hash) + HOST_NAME_FIELD_NUMBER; hash = (53 * hash) + getHostName().hashCode(); } if (hasPort()) { @@ -4862,7 +4862,7 @@ public final class HBaseProtos { hash = (53 * hash) + getPort(); } if (hasStartCode()) { - hash = (37 * hash) + STARTCODE_FIELD_NUMBER; + hash = (37 * hash) + START_CODE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getStartCode()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -5118,7 +5118,7 @@ public final class HBaseProtos { private int bitField0_; - // required string hostName = 1; + // required string host_name = 1; private java.lang.Object hostName_ = ""; public boolean hasHostName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -5175,7 +5175,7 @@ public final class HBaseProtos { return this; } - // optional uint64 startCode = 3; + // optional uint64 start_code = 3; private long startCode_ ; public boolean hasStartCode() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -7579,7 +7579,7 @@ public final class HBaseProtos { boolean hasTable(); String getTable(); - // optional int64 creationTime = 3 [default = 0]; + // optional int64 creation_time = 3 [default = 0]; boolean hasCreationTime(); long getCreationTime(); @@ -7753,8 +7753,8 @@ public final class HBaseProtos { } } - // optional int64 creationTime = 3 [default = 0]; - public static final int CREATIONTIME_FIELD_NUMBER = 3; + // optional int64 creation_time = 3 [default = 0]; + public static final int CREATION_TIME_FIELD_NUMBER = 3; private long creationTime_; public boolean hasCreationTime() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -7916,7 +7916,7 @@ public final class HBaseProtos { hash = (53 * hash) + getTable().hashCode(); } if (hasCreationTime()) { - hash = (37 * hash) + CREATIONTIME_FIELD_NUMBER; + hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCreationTime()); } if (hasType()) { @@ -8286,7 +8286,7 @@ public final class HBaseProtos { onChanged(); } - // optional int64 creationTime = 3 [default = 0]; + // optional int64 creation_time = 3 [default = 0]; private long creationTime_ ; public boolean hasCreationTime() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -8666,7 +8666,7 @@ public final class HBaseProtos { public interface LongMsgOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required int64 longMsg = 1; + // required int64 long_msg = 1; boolean hasLongMsg(); long getLongMsg(); } @@ -8699,8 +8699,8 @@ public final class HBaseProtos { } private int bitField0_; - // required int64 longMsg = 1; - public static final int LONGMSG_FIELD_NUMBER = 1; + // required int64 long_msg = 1; + public static final int LONG_MSG_FIELD_NUMBER = 1; private long longMsg_; public boolean hasLongMsg() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -8782,7 +8782,7 @@ public final class HBaseProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLongMsg()) { - hash = (37 * hash) + LONGMSG_FIELD_NUMBER; + hash = (37 * hash) + LONG_MSG_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLongMsg()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -9010,7 +9010,7 @@ public final class HBaseProtos { private int bitField0_; - // required int64 longMsg = 1; + // required int64 long_msg = 1; private long longMsg_ ; public boolean hasLongMsg() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9045,7 +9045,7 @@ public final class HBaseProtos { public interface BigDecimalMsgOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes bigdecimalMsg = 1; + // required bytes bigdecimal_msg = 1; boolean hasBigdecimalMsg(); com.google.protobuf.ByteString getBigdecimalMsg(); } @@ -9078,8 +9078,8 @@ public final class HBaseProtos { } private int bitField0_; - // required bytes bigdecimalMsg = 1; - public static final int BIGDECIMALMSG_FIELD_NUMBER = 1; + // required bytes bigdecimal_msg = 1; + public static final int BIGDECIMAL_MSG_FIELD_NUMBER = 1; private com.google.protobuf.ByteString bigdecimalMsg_; public boolean hasBigdecimalMsg() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9161,7 +9161,7 @@ public final class HBaseProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBigdecimalMsg()) { - hash = (37 * hash) + BIGDECIMALMSG_FIELD_NUMBER; + hash = (37 * hash) + BIGDECIMAL_MSG_FIELD_NUMBER; hash = (53 * hash) + getBigdecimalMsg().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -9389,7 +9389,7 @@ public final class HBaseProtos { private int bitField0_; - // required bytes bigdecimalMsg = 1; + // required bytes bigdecimal_msg = 1; private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY; public boolean hasBigdecimalMsg() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9427,11 +9427,11 @@ public final class HBaseProtos { public interface UUIDOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required uint64 leastSigBits = 1; + // required uint64 least_sig_bits = 1; boolean hasLeastSigBits(); long getLeastSigBits(); - // required uint64 mostSigBits = 2; + // required uint64 most_sig_bits = 2; boolean hasMostSigBits(); long getMostSigBits(); } @@ -9464,8 +9464,8 @@ public final class HBaseProtos { } private int bitField0_; - // required uint64 leastSigBits = 1; - public static final int LEASTSIGBITS_FIELD_NUMBER = 1; + // required uint64 least_sig_bits = 1; + public static final int LEAST_SIG_BITS_FIELD_NUMBER = 1; private long leastSigBits_; public boolean hasLeastSigBits() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9474,8 +9474,8 @@ public final class HBaseProtos { return leastSigBits_; } - // required uint64 mostSigBits = 2; - public static final int MOSTSIGBITS_FIELD_NUMBER = 2; + // required uint64 most_sig_bits = 2; + public static final int MOST_SIG_BITS_FIELD_NUMBER = 2; private long mostSigBits_; public boolean hasMostSigBits() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -9574,11 +9574,11 @@ public final class HBaseProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLeastSigBits()) { - hash = (37 * hash) + LEASTSIGBITS_FIELD_NUMBER; + hash = (37 * hash) + LEAST_SIG_BITS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLeastSigBits()); } if (hasMostSigBits()) { - hash = (37 * hash) + MOSTSIGBITS_FIELD_NUMBER; + hash = (37 * hash) + MOST_SIG_BITS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getMostSigBits()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -9824,7 +9824,7 @@ public final class HBaseProtos { private int bitField0_; - // required uint64 leastSigBits = 1; + // required uint64 least_sig_bits = 1; private long leastSigBits_ ; public boolean hasLeastSigBits() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9845,7 +9845,7 @@ public final class HBaseProtos { return this; } - // required uint64 mostSigBits = 2; + // required uint64 most_sig_bits = 2; private long mostSigBits_ ; public boolean hasMostSigBits() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -9971,42 +9971,43 @@ public final class HBaseProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\013hbase.proto\032\nCell.proto\"\225\001\n\013TableSchem" + + "\n\013hbase.proto\032\nCell.proto\"\226\001\n\013TableSchem" + "a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" + - "tesBytesPair\022+\n\016columnFamilies\030\003 \003(\0132\023.C" + - "olumnFamilySchema\022&\n\rconfiguration\030\004 \003(\013" + - "2\017.NameStringPair\"o\n\022ColumnFamilySchema\022" + - "\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byte" + - "sBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Name" + - "StringPair\"\207\001\n\nRegionInfo\022\020\n\010regionId\030\001 " + - "\002(\004\022\021\n\ttableName\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014" + - "\022\016\n\006endKey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl", - "it\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"0\n\014FavoredN" + - "odes\022 \n\013favoredNode\030\001 \003(\0132\013.ServerName\"\225" + - "\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$.Regio" + - "nSpecifier.RegionSpecifierType\022\r\n\005value\030" + - "\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REGION_N" + - "AME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tTimeRa" + - "nge\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"?\n\nServerN" + - "ame\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\021\n\ts" + - "tartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 " + - "\002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005v", - "alue\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(" + - "\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005fi" + - "rst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pa" + - "ir\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255\001\n\023Snap" + - "shotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002" + - " \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004type\030\004 " + - "\001(\0162\031.SnapshotDescription.Type:\005FLUSH\022\017\n" + - "\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" + - "FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n\007longMs" + - "g\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigdecimalMs", - "g\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001 \002(\004\022\023\n" + - "\013mostSigBits\030\002 \002(\004*r\n\013CompareType\022\010\n\004LES" + - "S\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT" + - "_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREATE" + - "R\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop.hbas" + - "e.protobuf.generatedB\013HBaseProtosH\001\240\001\001" + "tesBytesPair\022,\n\017column_families\030\003 \003(\0132\023." + + "ColumnFamilySchema\022&\n\rconfiguration\030\004 \003(" + + "\0132\017.NameStringPair\"o\n\022ColumnFamilySchema" + + "\022\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byt" + + "esBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Nam" + + "eStringPair\"\213\001\n\nRegionInfo\022\021\n\tregion_id\030" + + "\001 \002(\004\022\022\n\ntable_name\030\002 \002(\014\022\021\n\tstart_key\030\003" + + " \001(\014\022\017\n\007end_key\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r", + "\n\005split\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"1\n\014Fav" + + "oredNodes\022!\n\014favored_node\030\001 \003(\0132\013.Server" + + "Name\"\225\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$" + + ".RegionSpecifier.RegionSpecifierType\022\r\n\005" + + "value\030\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013RE" + + "GION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\t" + + "TimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"A\n\nS" + + "erverName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004port\030\002 \001" + + "(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n" + + "\004name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 ", + "\002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004n" + + "ame\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPa" + + "ir\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNam" + + "eInt64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"" + + "\256\001\n\023SnapshotDescription\022\014\n\004name\030\001 \002(\t\022\r\n" + + "\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022." + + "\n\004type\030\004 \001(\0162\031.SnapshotDescription.Type:" + + "\005FLUSH\022\017\n\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISAB" + + "LED\020\000\022\t\n\005FLUSH\020\001\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022" + + "\020\n\010long_msg\030\001 \002(\003\"\'\n\rBigDecimalMsg\022\026\n\016bi", + "gdecimal_msg\030\001 \002(\014\"5\n\004UUID\022\026\n\016least_sig_" + + "bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002 \002(\004*r\n\013Com" + + "pareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n" + + "\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQ" + + "UAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.apa" + + "che.hadoop.hbase.protobuf.generatedB\013HBa" + + "seProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HFileProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HFileProtos.java index d498d29..c1d240b 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HFileProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HFileProtos.java @@ -11,7 +11,7 @@ public final class HFileProtos { public interface FileInfoProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .BytesBytesPair mapEntry = 1; + // repeated .BytesBytesPair map_entry = 1; java.util.List getMapEntryList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index); @@ -49,8 +49,8 @@ public final class HFileProtos { return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable; } - // repeated .BytesBytesPair mapEntry = 1; - public static final int MAPENTRY_FIELD_NUMBER = 1; + // repeated .BytesBytesPair map_entry = 1; + public static final int MAP_ENTRY_FIELD_NUMBER = 1; private java.util.List mapEntry_; public java.util.List getMapEntryList() { return mapEntry_; @@ -142,7 +142,7 @@ public final class HFileProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getMapEntryCount() > 0) { - hash = (37 * hash) + MAPENTRY_FIELD_NUMBER; + hash = (37 * hash) + MAP_ENTRY_FIELD_NUMBER; hash = (53 * hash) + getMapEntryList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -404,7 +404,7 @@ public final class HFileProtos { private int bitField0_; - // repeated .BytesBytesPair mapEntry = 1; + // repeated .BytesBytesPair map_entry = 1; private java.util.List mapEntry_ = java.util.Collections.emptyList(); private void ensureMapEntryIsMutable() { @@ -604,51 +604,51 @@ public final class HFileProtos { public interface FileTrailerProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint64 fileInfoOffset = 1; + // optional uint64 file_info_offset = 1; boolean hasFileInfoOffset(); long getFileInfoOffset(); - // optional uint64 loadOnOpenDataOffset = 2; + // optional uint64 load_on_open_data_offset = 2; boolean hasLoadOnOpenDataOffset(); long getLoadOnOpenDataOffset(); - // optional uint64 uncompressedDataIndexSize = 3; + // optional uint64 uncompressed_data_index_size = 3; boolean hasUncompressedDataIndexSize(); long getUncompressedDataIndexSize(); - // optional uint64 totalUncompressedBytes = 4; + // optional uint64 total_uncompressed_bytes = 4; boolean hasTotalUncompressedBytes(); long getTotalUncompressedBytes(); - // optional uint32 dataIndexCount = 5; + // optional uint32 data_index_count = 5; boolean hasDataIndexCount(); int getDataIndexCount(); - // optional uint32 metaIndexCount = 6; + // optional uint32 meta_index_count = 6; boolean hasMetaIndexCount(); int getMetaIndexCount(); - // optional uint64 entryCount = 7; + // optional uint64 entry_count = 7; boolean hasEntryCount(); long getEntryCount(); - // optional uint32 numDataIndexLevels = 8; + // optional uint32 num_data_index_levels = 8; boolean hasNumDataIndexLevels(); int getNumDataIndexLevels(); - // optional uint64 firstDataBlockOffset = 9; + // optional uint64 first_data_block_offset = 9; boolean hasFirstDataBlockOffset(); long getFirstDataBlockOffset(); - // optional uint64 lastDataBlockOffset = 10; + // optional uint64 last_data_block_offset = 10; boolean hasLastDataBlockOffset(); long getLastDataBlockOffset(); - // optional string comparatorClassName = 11; + // optional string comparator_class_name = 11; boolean hasComparatorClassName(); String getComparatorClassName(); - // optional uint32 compressionCodec = 12; + // optional uint32 compression_codec = 12; boolean hasCompressionCodec(); int getCompressionCodec(); } @@ -681,8 +681,8 @@ public final class HFileProtos { } private int bitField0_; - // optional uint64 fileInfoOffset = 1; - public static final int FILEINFOOFFSET_FIELD_NUMBER = 1; + // optional uint64 file_info_offset = 1; + public static final int FILE_INFO_OFFSET_FIELD_NUMBER = 1; private long fileInfoOffset_; public boolean hasFileInfoOffset() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -691,8 +691,8 @@ public final class HFileProtos { return fileInfoOffset_; } - // optional uint64 loadOnOpenDataOffset = 2; - public static final int LOADONOPENDATAOFFSET_FIELD_NUMBER = 2; + // optional uint64 load_on_open_data_offset = 2; + public static final int LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER = 2; private long loadOnOpenDataOffset_; public boolean hasLoadOnOpenDataOffset() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -701,8 +701,8 @@ public final class HFileProtos { return loadOnOpenDataOffset_; } - // optional uint64 uncompressedDataIndexSize = 3; - public static final int UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER = 3; + // optional uint64 uncompressed_data_index_size = 3; + public static final int UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER = 3; private long uncompressedDataIndexSize_; public boolean hasUncompressedDataIndexSize() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -711,8 +711,8 @@ public final class HFileProtos { return uncompressedDataIndexSize_; } - // optional uint64 totalUncompressedBytes = 4; - public static final int TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER = 4; + // optional uint64 total_uncompressed_bytes = 4; + public static final int TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER = 4; private long totalUncompressedBytes_; public boolean hasTotalUncompressedBytes() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -721,8 +721,8 @@ public final class HFileProtos { return totalUncompressedBytes_; } - // optional uint32 dataIndexCount = 5; - public static final int DATAINDEXCOUNT_FIELD_NUMBER = 5; + // optional uint32 data_index_count = 5; + public static final int DATA_INDEX_COUNT_FIELD_NUMBER = 5; private int dataIndexCount_; public boolean hasDataIndexCount() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -731,8 +731,8 @@ public final class HFileProtos { return dataIndexCount_; } - // optional uint32 metaIndexCount = 6; - public static final int METAINDEXCOUNT_FIELD_NUMBER = 6; + // optional uint32 meta_index_count = 6; + public static final int META_INDEX_COUNT_FIELD_NUMBER = 6; private int metaIndexCount_; public boolean hasMetaIndexCount() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -741,8 +741,8 @@ public final class HFileProtos { return metaIndexCount_; } - // optional uint64 entryCount = 7; - public static final int ENTRYCOUNT_FIELD_NUMBER = 7; + // optional uint64 entry_count = 7; + public static final int ENTRY_COUNT_FIELD_NUMBER = 7; private long entryCount_; public boolean hasEntryCount() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -751,8 +751,8 @@ public final class HFileProtos { return entryCount_; } - // optional uint32 numDataIndexLevels = 8; - public static final int NUMDATAINDEXLEVELS_FIELD_NUMBER = 8; + // optional uint32 num_data_index_levels = 8; + public static final int NUM_DATA_INDEX_LEVELS_FIELD_NUMBER = 8; private int numDataIndexLevels_; public boolean hasNumDataIndexLevels() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -761,8 +761,8 @@ public final class HFileProtos { return numDataIndexLevels_; } - // optional uint64 firstDataBlockOffset = 9; - public static final int FIRSTDATABLOCKOFFSET_FIELD_NUMBER = 9; + // optional uint64 first_data_block_offset = 9; + public static final int FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 9; private long firstDataBlockOffset_; public boolean hasFirstDataBlockOffset() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -771,8 +771,8 @@ public final class HFileProtos { return firstDataBlockOffset_; } - // optional uint64 lastDataBlockOffset = 10; - public static final int LASTDATABLOCKOFFSET_FIELD_NUMBER = 10; + // optional uint64 last_data_block_offset = 10; + public static final int LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 10; private long lastDataBlockOffset_; public boolean hasLastDataBlockOffset() { return ((bitField0_ & 0x00000200) == 0x00000200); @@ -781,8 +781,8 @@ public final class HFileProtos { return lastDataBlockOffset_; } - // optional string comparatorClassName = 11; - public static final int COMPARATORCLASSNAME_FIELD_NUMBER = 11; + // optional string comparator_class_name = 11; + public static final int COMPARATOR_CLASS_NAME_FIELD_NUMBER = 11; private java.lang.Object comparatorClassName_; public boolean hasComparatorClassName() { return ((bitField0_ & 0x00000400) == 0x00000400); @@ -813,8 +813,8 @@ public final class HFileProtos { } } - // optional uint32 compressionCodec = 12; - public static final int COMPRESSIONCODEC_FIELD_NUMBER = 12; + // optional uint32 compression_codec = 12; + public static final int COMPRESSION_CODEC_FIELD_NUMBER = 12; private int compressionCodec_; public boolean hasCompressionCodec() { return ((bitField0_ & 0x00000800) == 0x00000800); @@ -1035,51 +1035,51 @@ public final class HFileProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasFileInfoOffset()) { - hash = (37 * hash) + FILEINFOOFFSET_FIELD_NUMBER; + hash = (37 * hash) + FILE_INFO_OFFSET_FIELD_NUMBER; hash = (53 * hash) + hashLong(getFileInfoOffset()); } if (hasLoadOnOpenDataOffset()) { - hash = (37 * hash) + LOADONOPENDATAOFFSET_FIELD_NUMBER; + hash = (37 * hash) + LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset()); } if (hasUncompressedDataIndexSize()) { - hash = (37 * hash) + UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER; + hash = (37 * hash) + UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getUncompressedDataIndexSize()); } if (hasTotalUncompressedBytes()) { - hash = (37 * hash) + TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER; + hash = (37 * hash) + TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER; hash = (53 * hash) + hashLong(getTotalUncompressedBytes()); } if (hasDataIndexCount()) { - hash = (37 * hash) + DATAINDEXCOUNT_FIELD_NUMBER; + hash = (37 * hash) + DATA_INDEX_COUNT_FIELD_NUMBER; hash = (53 * hash) + getDataIndexCount(); } if (hasMetaIndexCount()) { - hash = (37 * hash) + METAINDEXCOUNT_FIELD_NUMBER; + hash = (37 * hash) + META_INDEX_COUNT_FIELD_NUMBER; hash = (53 * hash) + getMetaIndexCount(); } if (hasEntryCount()) { - hash = (37 * hash) + ENTRYCOUNT_FIELD_NUMBER; + hash = (37 * hash) + ENTRY_COUNT_FIELD_NUMBER; hash = (53 * hash) + hashLong(getEntryCount()); } if (hasNumDataIndexLevels()) { - hash = (37 * hash) + NUMDATAINDEXLEVELS_FIELD_NUMBER; + hash = (37 * hash) + NUM_DATA_INDEX_LEVELS_FIELD_NUMBER; hash = (53 * hash) + getNumDataIndexLevels(); } if (hasFirstDataBlockOffset()) { - hash = (37 * hash) + FIRSTDATABLOCKOFFSET_FIELD_NUMBER; + hash = (37 * hash) + FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER; hash = (53 * hash) + hashLong(getFirstDataBlockOffset()); } if (hasLastDataBlockOffset()) { - hash = (37 * hash) + LASTDATABLOCKOFFSET_FIELD_NUMBER; + hash = (37 * hash) + LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLastDataBlockOffset()); } if (hasComparatorClassName()) { - hash = (37 * hash) + COMPARATORCLASSNAME_FIELD_NUMBER; + hash = (37 * hash) + COMPARATOR_CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getComparatorClassName().hashCode(); } if (hasCompressionCodec()) { - hash = (37 * hash) + COMPRESSIONCODEC_FIELD_NUMBER; + hash = (37 * hash) + COMPRESSION_CODEC_FIELD_NUMBER; hash = (53 * hash) + getCompressionCodec(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1457,7 +1457,7 @@ public final class HFileProtos { private int bitField0_; - // optional uint64 fileInfoOffset = 1; + // optional uint64 file_info_offset = 1; private long fileInfoOffset_ ; public boolean hasFileInfoOffset() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1478,7 +1478,7 @@ public final class HFileProtos { return this; } - // optional uint64 loadOnOpenDataOffset = 2; + // optional uint64 load_on_open_data_offset = 2; private long loadOnOpenDataOffset_ ; public boolean hasLoadOnOpenDataOffset() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1499,7 +1499,7 @@ public final class HFileProtos { return this; } - // optional uint64 uncompressedDataIndexSize = 3; + // optional uint64 uncompressed_data_index_size = 3; private long uncompressedDataIndexSize_ ; public boolean hasUncompressedDataIndexSize() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1520,7 +1520,7 @@ public final class HFileProtos { return this; } - // optional uint64 totalUncompressedBytes = 4; + // optional uint64 total_uncompressed_bytes = 4; private long totalUncompressedBytes_ ; public boolean hasTotalUncompressedBytes() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1541,7 +1541,7 @@ public final class HFileProtos { return this; } - // optional uint32 dataIndexCount = 5; + // optional uint32 data_index_count = 5; private int dataIndexCount_ ; public boolean hasDataIndexCount() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -1562,7 +1562,7 @@ public final class HFileProtos { return this; } - // optional uint32 metaIndexCount = 6; + // optional uint32 meta_index_count = 6; private int metaIndexCount_ ; public boolean hasMetaIndexCount() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -1583,7 +1583,7 @@ public final class HFileProtos { return this; } - // optional uint64 entryCount = 7; + // optional uint64 entry_count = 7; private long entryCount_ ; public boolean hasEntryCount() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -1604,7 +1604,7 @@ public final class HFileProtos { return this; } - // optional uint32 numDataIndexLevels = 8; + // optional uint32 num_data_index_levels = 8; private int numDataIndexLevels_ ; public boolean hasNumDataIndexLevels() { return ((bitField0_ & 0x00000080) == 0x00000080); @@ -1625,7 +1625,7 @@ public final class HFileProtos { return this; } - // optional uint64 firstDataBlockOffset = 9; + // optional uint64 first_data_block_offset = 9; private long firstDataBlockOffset_ ; public boolean hasFirstDataBlockOffset() { return ((bitField0_ & 0x00000100) == 0x00000100); @@ -1646,7 +1646,7 @@ public final class HFileProtos { return this; } - // optional uint64 lastDataBlockOffset = 10; + // optional uint64 last_data_block_offset = 10; private long lastDataBlockOffset_ ; public boolean hasLastDataBlockOffset() { return ((bitField0_ & 0x00000200) == 0x00000200); @@ -1667,7 +1667,7 @@ public final class HFileProtos { return this; } - // optional string comparatorClassName = 11; + // optional string comparator_class_name = 11; private java.lang.Object comparatorClassName_ = ""; public boolean hasComparatorClassName() { return ((bitField0_ & 0x00000400) == 0x00000400); @@ -1703,7 +1703,7 @@ public final class HFileProtos { onChanged(); } - // optional uint32 compressionCodec = 12; + // optional uint32 compression_codec = 12; private int compressionCodec_ ; public boolean hasCompressionCodec() { return ((bitField0_ & 0x00000800) == 0x00000800); @@ -1754,19 +1754,20 @@ public final class HFileProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\013HFile.proto\032\013hbase.proto\"2\n\rFileInfoPr" + - "oto\022!\n\010mapEntry\030\001 \003(\0132\017.BytesBytesPair\"\335" + - "\002\n\020FileTrailerProto\022\026\n\016fileInfoOffset\030\001 " + - "\001(\004\022\034\n\024loadOnOpenDataOffset\030\002 \001(\004\022!\n\031unc" + - "ompressedDataIndexSize\030\003 \001(\004\022\036\n\026totalUnc" + - "ompressedBytes\030\004 \001(\004\022\026\n\016dataIndexCount\030\005" + - " \001(\r\022\026\n\016metaIndexCount\030\006 \001(\r\022\022\n\nentryCou" + - "nt\030\007 \001(\004\022\032\n\022numDataIndexLevels\030\010 \001(\r\022\034\n\024" + - "firstDataBlockOffset\030\t \001(\004\022\033\n\023lastDataBl" + - "ockOffset\030\n \001(\004\022\033\n\023comparatorClassName\030\013", - " \001(\t\022\030\n\020compressionCodec\030\014 \001(\rBA\n*org.ap" + - "ache.hadoop.hbase.protobuf.generatedB\013HF" + - "ileProtosH\001\210\001\001\240\001\001" + "\n\013HFile.proto\032\013hbase.proto\"3\n\rFileInfoPr" + + "oto\022\"\n\tmap_entry\030\001 \003(\0132\017.BytesBytesPair\"" + + "\371\002\n\020FileTrailerProto\022\030\n\020file_info_offset" + + "\030\001 \001(\004\022 \n\030load_on_open_data_offset\030\002 \001(\004" + + "\022$\n\034uncompressed_data_index_size\030\003 \001(\004\022 " + + "\n\030total_uncompressed_bytes\030\004 \001(\004\022\030\n\020data" + + "_index_count\030\005 \001(\r\022\030\n\020meta_index_count\030\006" + + " \001(\r\022\023\n\013entry_count\030\007 \001(\004\022\035\n\025num_data_in" + + "dex_levels\030\010 \001(\r\022\037\n\027first_data_block_off" + + "set\030\t \001(\004\022\036\n\026last_data_block_offset\030\n \001(", + "\004\022\035\n\025comparator_class_name\030\013 \001(\t\022\031\n\021comp" + + "ression_codec\030\014 \001(\rBA\n*org.apache.hadoop" + + ".hbase.protobuf.generatedB\013HFileProtosH\001" + + "\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/LoadBalancerProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/LoadBalancerProtos.java index 48243e6..937c6d2 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/LoadBalancerProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/LoadBalancerProtos.java @@ -11,7 +11,7 @@ public final class LoadBalancerProtos { public interface LoadBalancerStateOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bool balancerOn = 1; + // optional bool balancer_on = 1; boolean hasBalancerOn(); boolean getBalancerOn(); } @@ -44,8 +44,8 @@ public final class LoadBalancerProtos { } private int bitField0_; - // optional bool balancerOn = 1; - public static final int BALANCERON_FIELD_NUMBER = 1; + // optional bool balancer_on = 1; + public static final int BALANCER_ON_FIELD_NUMBER = 1; private boolean balancerOn_; public boolean hasBalancerOn() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -123,7 +123,7 @@ public final class LoadBalancerProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBalancerOn()) { - hash = (37 * hash) + BALANCERON_FIELD_NUMBER; + hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getBalancerOn()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -347,7 +347,7 @@ public final class LoadBalancerProtos { private int bitField0_; - // optional bool balancerOn = 1; + // optional bool balancer_on = 1; private boolean balancerOn_ ; public boolean hasBalancerOn() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -393,10 +393,10 @@ public final class LoadBalancerProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\022LoadBalancer.proto\"\'\n\021LoadBalancerStat" + - "e\022\022\n\nbalancerOn\030\001 \001(\010BE\n*org.apache.hado" + - "op.hbase.protobuf.generatedB\022LoadBalance" + - "rProtosH\001\240\001\001" + "\n\022LoadBalancer.proto\"(\n\021LoadBalancerStat" + + "e\022\023\n\013balancer_on\030\001 \001(\010BE\n*org.apache.had" + + "oop.hbase.protobuf.generatedB\022LoadBalanc" + + "erProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java index f2c6ed3..7f7cc70 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java @@ -11,11 +11,11 @@ public final class MasterAdminProtos { public interface AddColumnRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // required .ColumnFamilySchema columnFamilies = 2; + // required .ColumnFamilySchema column_families = 2; boolean hasColumnFamilies(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(); @@ -49,8 +49,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -59,8 +59,8 @@ public final class MasterAdminProtos { return tableName_; } - // required .ColumnFamilySchema columnFamilies = 2; - public static final int COLUMNFAMILIES_FIELD_NUMBER = 2; + // required .ColumnFamilySchema column_families = 2; + public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_; public boolean hasColumnFamilies() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -166,11 +166,11 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasColumnFamilies()) { - hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER; hash = (53 * hash) + getColumnFamilies().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -433,7 +433,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -457,7 +457,7 @@ public final class MasterAdminProtos { return this; } - // required .ColumnFamilySchema columnFamilies = 2; + // required .ColumnFamilySchema column_families = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; @@ -861,11 +861,11 @@ public final class MasterAdminProtos { public interface DeleteColumnRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // required bytes columnName = 2; + // required bytes column_name = 2; boolean hasColumnName(); com.google.protobuf.ByteString getColumnName(); } @@ -898,8 +898,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -908,8 +908,8 @@ public final class MasterAdminProtos { return tableName_; } - // required bytes columnName = 2; - public static final int COLUMNNAME_FIELD_NUMBER = 2; + // required bytes column_name = 2; + public static final int COLUMN_NAME_FIELD_NUMBER = 2; private com.google.protobuf.ByteString columnName_; public boolean hasColumnName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1008,11 +1008,11 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasColumnName()) { - hash = (37 * hash) + COLUMNNAME_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER; hash = (53 * hash) + getColumnName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1258,7 +1258,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1282,7 +1282,7 @@ public final class MasterAdminProtos { return this; } - // required bytes columnName = 2; + // required bytes column_name = 2; private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasColumnName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1620,11 +1620,11 @@ public final class MasterAdminProtos { public interface ModifyColumnRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // required .ColumnFamilySchema columnFamilies = 2; + // required .ColumnFamilySchema column_families = 2; boolean hasColumnFamilies(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(); @@ -1658,8 +1658,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1668,8 +1668,8 @@ public final class MasterAdminProtos { return tableName_; } - // required .ColumnFamilySchema columnFamilies = 2; - public static final int COLUMNFAMILIES_FIELD_NUMBER = 2; + // required .ColumnFamilySchema column_families = 2; + public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_; public boolean hasColumnFamilies() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1775,11 +1775,11 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasColumnFamilies()) { - hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; + hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER; hash = (53 * hash) + getColumnFamilies().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2042,7 +2042,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2066,7 +2066,7 @@ public final class MasterAdminProtos { return this; } - // required .ColumnFamilySchema columnFamilies = 2; + // required .ColumnFamilySchema column_families = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; @@ -2475,7 +2475,7 @@ public final class MasterAdminProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); - // optional .ServerName destServerName = 2; + // optional .ServerName dest_server_name = 2; boolean hasDestServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder(); @@ -2522,8 +2522,8 @@ public final class MasterAdminProtos { return region_; } - // optional .ServerName destServerName = 2; - public static final int DESTSERVERNAME_FIELD_NUMBER = 2; + // optional .ServerName dest_server_name = 2; + public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_; public boolean hasDestServerName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2635,7 +2635,7 @@ public final class MasterAdminProtos { hash = (53 * hash) + getRegion().hashCode(); } if (hasDestServerName()) { - hash = (37 * hash) + DESTSERVERNAME_FIELD_NUMBER; + hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getDestServerName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3003,7 +3003,7 @@ public final class MasterAdminProtos { return regionBuilder_; } - // optional .ServerName destServerName = 2; + // optional .ServerName dest_server_name = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_; @@ -3407,12 +3407,12 @@ public final class MasterAdminProtos { public interface DispatchMergingRegionsRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .RegionSpecifier regionA = 1; + // required .RegionSpecifier region_a = 1; boolean hasRegionA(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); - // required .RegionSpecifier regionB = 2; + // required .RegionSpecifier region_b = 2; boolean hasRegionB(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); @@ -3450,8 +3450,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required .RegionSpecifier regionA = 1; - public static final int REGIONA_FIELD_NUMBER = 1; + // required .RegionSpecifier region_a = 1; + public static final int REGION_A_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; public boolean hasRegionA() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3463,8 +3463,8 @@ public final class MasterAdminProtos { return regionA_; } - // required .RegionSpecifier regionB = 2; - public static final int REGIONB_FIELD_NUMBER = 2; + // required .RegionSpecifier region_b = 2; + public static final int REGION_B_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; public boolean hasRegionB() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -3597,11 +3597,11 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionA()) { - hash = (37 * hash) + REGIONA_FIELD_NUMBER; + hash = (37 * hash) + REGION_A_FIELD_NUMBER; hash = (53 * hash) + getRegionA().hashCode(); } if (hasRegionB()) { - hash = (37 * hash) + REGIONB_FIELD_NUMBER; + hash = (37 * hash) + REGION_B_FIELD_NUMBER; hash = (53 * hash) + getRegionB().hashCode(); } if (hasForcible()) { @@ -3899,7 +3899,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required .RegionSpecifier regionA = 1; + // required .RegionSpecifier region_a = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; @@ -3989,7 +3989,7 @@ public final class MasterAdminProtos { return regionABuilder_; } - // required .RegionSpecifier regionB = 2; + // required .RegionSpecifier region_b = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; @@ -6799,12 +6799,12 @@ public final class MasterAdminProtos { public interface CreateTableRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .TableSchema tableSchema = 1; + // required .TableSchema table_schema = 1; boolean hasTableSchema(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder(); - // repeated bytes splitKeys = 2; + // repeated bytes split_keys = 2; java.util.List getSplitKeysList(); int getSplitKeysCount(); com.google.protobuf.ByteString getSplitKeys(int index); @@ -6838,8 +6838,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required .TableSchema tableSchema = 1; - public static final int TABLESCHEMA_FIELD_NUMBER = 1; + // required .TableSchema table_schema = 1; + public static final int TABLE_SCHEMA_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_; public boolean hasTableSchema() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6851,8 +6851,8 @@ public final class MasterAdminProtos { return tableSchema_; } - // repeated bytes splitKeys = 2; - public static final int SPLITKEYS_FIELD_NUMBER = 2; + // repeated bytes split_keys = 2; + public static final int SPLIT_KEYS_FIELD_NUMBER = 2; private java.util.List splitKeys_; public java.util.List getSplitKeysList() { @@ -6957,11 +6957,11 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableSchema()) { - hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; + hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchema().hashCode(); } if (getSplitKeysCount() > 0) { - hash = (37 * hash) + SPLITKEYS_FIELD_NUMBER; + hash = (37 * hash) + SPLIT_KEYS_FIELD_NUMBER; hash = (53 * hash) + getSplitKeysList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -7228,7 +7228,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required .TableSchema tableSchema = 1; + // required .TableSchema table_schema = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; @@ -7318,7 +7318,7 @@ public final class MasterAdminProtos { return tableSchemaBuilder_; } - // repeated bytes splitKeys = 2; + // repeated bytes split_keys = 2; private java.util.List splitKeys_ = java.util.Collections.emptyList();; private void ensureSplitKeysIsMutable() { if (!((bitField0_ & 0x00000002) == 0x00000002)) { @@ -7683,7 +7683,7 @@ public final class MasterAdminProtos { public interface DeleteTableRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); } @@ -7716,8 +7716,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -7799,7 +7799,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -8027,7 +8027,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -8365,7 +8365,7 @@ public final class MasterAdminProtos { public interface EnableTableRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); } @@ -8398,8 +8398,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -8481,7 +8481,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -8709,7 +8709,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9047,7 +9047,7 @@ public final class MasterAdminProtos { public interface DisableTableRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); } @@ -9080,8 +9080,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9163,7 +9163,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -9391,7 +9391,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9729,11 +9729,11 @@ public final class MasterAdminProtos { public interface ModifyTableRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // required .TableSchema tableSchema = 2; + // required .TableSchema table_schema = 2; boolean hasTableSchema(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder(); @@ -9767,8 +9767,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -9777,8 +9777,8 @@ public final class MasterAdminProtos { return tableName_; } - // required .TableSchema tableSchema = 2; - public static final int TABLESCHEMA_FIELD_NUMBER = 2; + // required .TableSchema table_schema = 2; + public static final int TABLE_SCHEMA_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_; public boolean hasTableSchema() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -9884,11 +9884,11 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasTableSchema()) { - hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; + hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchema().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -10151,7 +10151,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -10175,7 +10175,7 @@ public final class MasterAdminProtos { return this; } - // required .TableSchema tableSchema = 2; + // required .TableSchema table_schema = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; @@ -12079,7 +12079,7 @@ public final class MasterAdminProtos { public interface BalanceResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bool balancerRan = 1; + // required bool balancer_ran = 1; boolean hasBalancerRan(); boolean getBalancerRan(); } @@ -12112,8 +12112,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required bool balancerRan = 1; - public static final int BALANCERRAN_FIELD_NUMBER = 1; + // required bool balancer_ran = 1; + public static final int BALANCER_RAN_FIELD_NUMBER = 1; private boolean balancerRan_; public boolean hasBalancerRan() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -12195,7 +12195,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBalancerRan()) { - hash = (37 * hash) + BALANCERRAN_FIELD_NUMBER; + hash = (37 * hash) + BALANCER_RAN_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getBalancerRan()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -12423,7 +12423,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required bool balancerRan = 1; + // required bool balancer_ran = 1; private boolean balancerRan_ ; public boolean hasBalancerRan() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -12903,7 +12903,7 @@ public final class MasterAdminProtos { public interface SetBalancerRunningResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bool prevBalanceValue = 1; + // optional bool prev_balance_value = 1; boolean hasPrevBalanceValue(); boolean getPrevBalanceValue(); } @@ -12936,8 +12936,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // optional bool prevBalanceValue = 1; - public static final int PREVBALANCEVALUE_FIELD_NUMBER = 1; + // optional bool prev_balance_value = 1; + public static final int PREV_BALANCE_VALUE_FIELD_NUMBER = 1; private boolean prevBalanceValue_; public boolean hasPrevBalanceValue() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -13015,7 +13015,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasPrevBalanceValue()) { - hash = (37 * hash) + PREVBALANCEVALUE_FIELD_NUMBER; + hash = (37 * hash) + PREV_BALANCE_VALUE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getPrevBalanceValue()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -13239,7 +13239,7 @@ public final class MasterAdminProtos { private int bitField0_; - // optional bool prevBalanceValue = 1; + // optional bool prev_balance_value = 1; private boolean prevBalanceValue_ ; public boolean hasPrevBalanceValue() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -13574,7 +13574,7 @@ public final class MasterAdminProtos { public interface CatalogScanResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional int32 scanResult = 1; + // optional int32 scan_result = 1; boolean hasScanResult(); int getScanResult(); } @@ -13607,8 +13607,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // optional int32 scanResult = 1; - public static final int SCANRESULT_FIELD_NUMBER = 1; + // optional int32 scan_result = 1; + public static final int SCAN_RESULT_FIELD_NUMBER = 1; private int scanResult_; public boolean hasScanResult() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -13686,7 +13686,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasScanResult()) { - hash = (37 * hash) + SCANRESULT_FIELD_NUMBER; + hash = (37 * hash) + SCAN_RESULT_FIELD_NUMBER; hash = (53 * hash) + getScanResult(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -13910,7 +13910,7 @@ public final class MasterAdminProtos { private int bitField0_; - // optional int32 scanResult = 1; + // optional int32 scan_result = 1; private int scanResult_ ; public boolean hasScanResult() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -14324,7 +14324,7 @@ public final class MasterAdminProtos { public interface EnableCatalogJanitorResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bool prevValue = 1; + // optional bool prev_value = 1; boolean hasPrevValue(); boolean getPrevValue(); } @@ -14357,8 +14357,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // optional bool prevValue = 1; - public static final int PREVVALUE_FIELD_NUMBER = 1; + // optional bool prev_value = 1; + public static final int PREV_VALUE_FIELD_NUMBER = 1; private boolean prevValue_; public boolean hasPrevValue() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -14436,7 +14436,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasPrevValue()) { - hash = (37 * hash) + PREVVALUE_FIELD_NUMBER; + hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getPrevValue()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -14660,7 +14660,7 @@ public final class MasterAdminProtos { private int bitField0_; - // optional bool prevValue = 1; + // optional bool prev_value = 1; private boolean prevValue_ ; public boolean hasPrevValue() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -15847,7 +15847,7 @@ public final class MasterAdminProtos { public interface TakeSnapshotResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required int64 expectedTimeout = 1; + // required int64 expected_timeout = 1; boolean hasExpectedTimeout(); long getExpectedTimeout(); } @@ -15880,8 +15880,8 @@ public final class MasterAdminProtos { } private int bitField0_; - // required int64 expectedTimeout = 1; - public static final int EXPECTEDTIMEOUT_FIELD_NUMBER = 1; + // required int64 expected_timeout = 1; + public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1; private long expectedTimeout_; public boolean hasExpectedTimeout() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -15963,7 +15963,7 @@ public final class MasterAdminProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasExpectedTimeout()) { - hash = (37 * hash) + EXPECTEDTIMEOUT_FIELD_NUMBER; + hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER; hash = (53 * hash) + hashLong(getExpectedTimeout()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -16191,7 +16191,7 @@ public final class MasterAdminProtos { private int bitField0_; - // required int64 expectedTimeout = 1; + // required int64 expected_timeout = 1; private long expectedTimeout_ ; public boolean hasExpectedTimeout() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -22752,116 +22752,116 @@ public final class MasterAdminProtos { static { java.lang.String[] descriptorData = { "\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" + - ".proto\032\014Client.proto\"R\n\020AddColumnRequest" + - "\022\021\n\ttableName\030\001 \002(\014\022+\n\016columnFamilies\030\002 " + - "\002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnRes" + - "ponse\"<\n\023DeleteColumnRequest\022\021\n\ttableNam" + - "e\030\001 \002(\014\022\022\n\ncolumnName\030\002 \002(\014\"\026\n\024DeleteCol" + - "umnResponse\"U\n\023ModifyColumnRequest\022\021\n\tta" + - "bleName\030\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023." + - "ColumnFamilySchema\"\026\n\024ModifyColumnRespon" + - "se\"Z\n\021MoveRegionRequest\022 \n\006region\030\001 \002(\0132", - "\020.RegionSpecifier\022#\n\016destServerName\030\002 \001(" + - "\0132\013.ServerName\"\024\n\022MoveRegionResponse\"~\n\035" + - "DispatchMergingRegionsRequest\022!\n\007regionA" + - "\030\001 \002(\0132\020.RegionSpecifier\022!\n\007regionB\030\002 \002(" + - "\0132\020.RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005f" + - "alse\" \n\036DispatchMergingRegionsResponse\"7" + - "\n\023AssignRegionRequest\022 \n\006region\030\001 \002(\0132\020." + - "RegionSpecifier\"\026\n\024AssignRegionResponse\"" + - "O\n\025UnassignRegionRequest\022 \n\006region\030\001 \002(\013" + - "2\020.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005false", - "\"\030\n\026UnassignRegionResponse\"8\n\024OfflineReg" + - "ionRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpeci" + - "fier\"\027\n\025OfflineRegionResponse\"J\n\022CreateT" + - "ableRequest\022!\n\013tableSchema\030\001 \002(\0132\014.Table" + - "Schema\022\021\n\tsplitKeys\030\002 \003(\014\"\025\n\023CreateTable" + - "Response\"\'\n\022DeleteTableRequest\022\021\n\ttableN" + - "ame\030\001 \002(\014\"\025\n\023DeleteTableResponse\"\'\n\022Enab" + - "leTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023Ena" + - "bleTableResponse\"(\n\023DisableTableRequest\022" + - "\021\n\ttableName\030\001 \002(\014\"\026\n\024DisableTableRespon", - "se\"J\n\022ModifyTableRequest\022\021\n\ttableName\030\001 " + - "\002(\014\022!\n\013tableSchema\030\002 \002(\0132\014.TableSchema\"\025" + - "\n\023ModifyTableResponse\"\021\n\017ShutdownRequest" + - "\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReques" + - "t\"\024\n\022StopMasterResponse\"\020\n\016BalanceReques" + - "t\"&\n\017BalanceResponse\022\023\n\013balancerRan\030\001 \002(" + - "\010\"<\n\031SetBalancerRunningRequest\022\n\n\002on\030\001 \002" + - "(\010\022\023\n\013synchronous\030\002 \001(\010\"6\n\032SetBalancerRu" + - "nningResponse\022\030\n\020prevBalanceValue\030\001 \001(\010\"" + - "\024\n\022CatalogScanRequest\")\n\023CatalogScanResp", - "onse\022\022\n\nscanResult\030\001 \001(\005\"-\n\033EnableCatalo" + - "gJanitorRequest\022\016\n\006enable\030\001 \002(\010\"1\n\034Enabl" + - "eCatalogJanitorResponse\022\021\n\tprevValue\030\001 \001" + - "(\010\" \n\036IsCatalogJanitorEnabledRequest\"0\n\037" + - "IsCatalogJanitorEnabledResponse\022\r\n\005value" + - "\030\001 \002(\010\"=\n\023TakeSnapshotRequest\022&\n\010snapsho" + - "t\030\001 \002(\0132\024.SnapshotDescription\"/\n\024TakeSna" + - "pshotResponse\022\027\n\017expectedTimeout\030\001 \002(\003\"\025" + - "\n\023ListSnapshotRequest\"?\n\024ListSnapshotRes" + - "ponse\022\'\n\tsnapshots\030\001 \003(\0132\024.SnapshotDescr", - "iption\"?\n\025DeleteSnapshotRequest\022&\n\010snaps" + - "hot\030\001 \002(\0132\024.SnapshotDescription\"\030\n\026Delet" + - "eSnapshotResponse\"@\n\026RestoreSnapshotRequ" + - "est\022&\n\010snapshot\030\001 \002(\0132\024.SnapshotDescript" + - "ion\"\031\n\027RestoreSnapshotResponse\"?\n\025IsSnap" + - "shotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snap" + - "shotDescription\"U\n\026IsSnapshotDoneRespons" + - "e\022\023\n\004done\030\001 \001(\010:\005false\022&\n\010snapshot\030\002 \001(\013" + - "2\024.SnapshotDescription\"F\n\034IsRestoreSnaps" + - "hotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snaps", - "hotDescription\"3\n\035IsRestoreSnapshotDoneR" + - "esponse\022\022\n\004done\030\001 \001(\010:\004true2\305\016\n\022MasterAd" + - "minService\0222\n\taddColumn\022\021.AddColumnReque" + - "st\032\022.AddColumnResponse\022;\n\014deleteColumn\022\024" + - ".DeleteColumnRequest\032\025.DeleteColumnRespo" + - "nse\022;\n\014modifyColumn\022\024.ModifyColumnReques" + - "t\032\025.ModifyColumnResponse\0225\n\nmoveRegion\022\022" + - ".MoveRegionRequest\032\023.MoveRegionResponse\022" + - "Y\n\026dispatchMergingRegions\022\036.DispatchMerg" + - "ingRegionsRequest\032\037.DispatchMergingRegio", - "nsResponse\022;\n\014assignRegion\022\024.AssignRegio" + - "nRequest\032\025.AssignRegionResponse\022A\n\016unass" + - "ignRegion\022\026.UnassignRegionRequest\032\027.Unas" + - "signRegionResponse\022>\n\rofflineRegion\022\025.Of" + - "flineRegionRequest\032\026.OfflineRegionRespon" + - "se\0228\n\013deleteTable\022\023.DeleteTableRequest\032\024" + - ".DeleteTableResponse\0228\n\013enableTable\022\023.En" + - "ableTableRequest\032\024.EnableTableResponse\022;" + - "\n\014disableTable\022\024.DisableTableRequest\032\025.D" + - "isableTableResponse\0228\n\013modifyTable\022\023.Mod", - "ifyTableRequest\032\024.ModifyTableResponse\0228\n" + - "\013createTable\022\023.CreateTableRequest\032\024.Crea" + - "teTableResponse\022/\n\010shutdown\022\020.ShutdownRe" + - "quest\032\021.ShutdownResponse\0225\n\nstopMaster\022\022" + - ".StopMasterRequest\032\023.StopMasterResponse\022" + - ",\n\007balance\022\017.BalanceRequest\032\020.BalanceRes" + - "ponse\022M\n\022setBalancerRunning\022\032.SetBalance" + - "rRunningRequest\032\033.SetBalancerRunningResp" + - "onse\022;\n\016runCatalogScan\022\023.CatalogScanRequ" + - "est\032\024.CatalogScanResponse\022S\n\024enableCatal", - "ogJanitor\022\034.EnableCatalogJanitorRequest\032" + - "\035.EnableCatalogJanitorResponse\022\\\n\027isCata" + - "logJanitorEnabled\022\037.IsCatalogJanitorEnab" + - "ledRequest\032 .IsCatalogJanitorEnabledResp" + - "onse\022L\n\021execMasterService\022\032.CoprocessorS" + - "erviceRequest\032\033.CoprocessorServiceRespon" + - "se\0227\n\010snapshot\022\024.TakeSnapshotRequest\032\025.T" + - "akeSnapshotResponse\022D\n\025getCompletedSnaps" + - "hots\022\024.ListSnapshotRequest\032\025.ListSnapsho" + - "tResponse\022A\n\016deleteSnapshot\022\026.DeleteSnap", - "shotRequest\032\027.DeleteSnapshotResponse\022A\n\016" + - "isSnapshotDone\022\026.IsSnapshotDoneRequest\032\027" + - ".IsSnapshotDoneResponse\022D\n\017restoreSnapsh" + - "ot\022\027.RestoreSnapshotRequest\032\030.RestoreSna" + - "pshotResponse\022V\n\025isRestoreSnapshotDone\022\035" + - ".IsRestoreSnapshotDoneRequest\032\036.IsRestor" + - "eSnapshotDoneResponse\022D\n\017isMasterRunning" + - "\022\027.IsMasterRunningRequest\032\030.IsMasterRunn" + - "ingResponseBG\n*org.apache.hadoop.hbase.p" + - "rotobuf.generatedB\021MasterAdminProtosH\001\210\001", - "\001\240\001\001" + ".proto\032\014Client.proto\"T\n\020AddColumnRequest" + + "\022\022\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030" + + "\002 \002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnR" + + "esponse\">\n\023DeleteColumnRequest\022\022\n\ntable_" + + "name\030\001 \002(\014\022\023\n\013column_name\030\002 \002(\014\"\026\n\024Delet" + + "eColumnResponse\"W\n\023ModifyColumnRequest\022\022" + + "\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030\002 " + + "\002(\0132\023.ColumnFamilySchema\"\026\n\024ModifyColumn" + + "Response\"\\\n\021MoveRegionRequest\022 \n\006region\030", + "\001 \002(\0132\020.RegionSpecifier\022%\n\020dest_server_n" + + "ame\030\002 \001(\0132\013.ServerName\"\024\n\022MoveRegionResp" + + "onse\"\200\001\n\035DispatchMergingRegionsRequest\022\"" + + "\n\010region_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010re" + + "gion_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcib" + + "le\030\003 \001(\010:\005false\" \n\036DispatchMergingRegion" + + "sResponse\"7\n\023AssignRegionRequest\022 \n\006regi" + + "on\030\001 \002(\0132\020.RegionSpecifier\"\026\n\024AssignRegi" + + "onResponse\"O\n\025UnassignRegionRequest\022 \n\006r" + + "egion\030\001 \002(\0132\020.RegionSpecifier\022\024\n\005force\030\002", + " \001(\010:\005false\"\030\n\026UnassignRegionResponse\"8\n" + + "\024OfflineRegionRequest\022 \n\006region\030\001 \002(\0132\020." + + "RegionSpecifier\"\027\n\025OfflineRegionResponse" + + "\"L\n\022CreateTableRequest\022\"\n\014table_schema\030\001" + + " \002(\0132\014.TableSchema\022\022\n\nsplit_keys\030\002 \003(\014\"\025" + + "\n\023CreateTableResponse\"(\n\022DeleteTableRequ" + + "est\022\022\n\ntable_name\030\001 \002(\014\"\025\n\023DeleteTableRe" + + "sponse\"(\n\022EnableTableRequest\022\022\n\ntable_na" + + "me\030\001 \002(\014\"\025\n\023EnableTableResponse\")\n\023Disab" + + "leTableRequest\022\022\n\ntable_name\030\001 \002(\014\"\026\n\024Di", + "sableTableResponse\"L\n\022ModifyTableRequest" + + "\022\022\n\ntable_name\030\001 \002(\014\022\"\n\014table_schema\030\002 \002" + + "(\0132\014.TableSchema\"\025\n\023ModifyTableResponse\"" + + "\021\n\017ShutdownRequest\"\022\n\020ShutdownResponse\"\023" + + "\n\021StopMasterRequest\"\024\n\022StopMasterRespons" + + "e\"\020\n\016BalanceRequest\"\'\n\017BalanceResponse\022\024" + + "\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalancerRunni" + + "ngRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001" + + "(\010\"8\n\032SetBalancerRunningResponse\022\032\n\022prev" + + "_balance_value\030\001 \001(\010\"\024\n\022CatalogScanReque", + "st\"*\n\023CatalogScanResponse\022\023\n\013scan_result" + + "\030\001 \001(\005\"-\n\033EnableCatalogJanitorRequest\022\016\n" + + "\006enable\030\001 \002(\010\"2\n\034EnableCatalogJanitorRes" + + "ponse\022\022\n\nprev_value\030\001 \001(\010\" \n\036IsCatalogJa" + + "nitorEnabledRequest\"0\n\037IsCatalogJanitorE" + + "nabledResponse\022\r\n\005value\030\001 \002(\010\"=\n\023TakeSna" + + "pshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snapsho" + + "tDescription\"0\n\024TakeSnapshotResponse\022\030\n\020" + + "expected_timeout\030\001 \002(\003\"\025\n\023ListSnapshotRe" + + "quest\"?\n\024ListSnapshotResponse\022\'\n\tsnapsho", + "ts\030\001 \003(\0132\024.SnapshotDescription\"?\n\025Delete" + + "SnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snap" + + "shotDescription\"\030\n\026DeleteSnapshotRespons" + + "e\"@\n\026RestoreSnapshotRequest\022&\n\010snapshot\030" + + "\001 \002(\0132\024.SnapshotDescription\"\031\n\027RestoreSn" + + "apshotResponse\"?\n\025IsSnapshotDoneRequest\022" + + "&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"" + + "U\n\026IsSnapshotDoneResponse\022\023\n\004done\030\001 \001(\010:" + + "\005false\022&\n\010snapshot\030\002 \001(\0132\024.SnapshotDescr" + + "iption\"F\n\034IsRestoreSnapshotDoneRequest\022&", + "\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"3" + + "\n\035IsRestoreSnapshotDoneResponse\022\022\n\004done\030" + + "\001 \001(\010:\004true2\305\016\n\022MasterAdminService\0222\n\tAd" + + "dColumn\022\021.AddColumnRequest\032\022.AddColumnRe" + + "sponse\022;\n\014DeleteColumn\022\024.DeleteColumnReq" + + "uest\032\025.DeleteColumnResponse\022;\n\014ModifyCol" + + "umn\022\024.ModifyColumnRequest\032\025.ModifyColumn" + + "Response\0225\n\nMoveRegion\022\022.MoveRegionReque" + + "st\032\023.MoveRegionResponse\022Y\n\026DispatchMergi" + + "ngRegions\022\036.DispatchMergingRegionsReques", + "t\032\037.DispatchMergingRegionsResponse\022;\n\014As" + + "signRegion\022\024.AssignRegionRequest\032\025.Assig" + + "nRegionResponse\022A\n\016UnassignRegion\022\026.Unas" + + "signRegionRequest\032\027.UnassignRegionRespon" + + "se\022>\n\rOfflineRegion\022\025.OfflineRegionReque" + + "st\032\026.OfflineRegionResponse\0228\n\013DeleteTabl" + + "e\022\023.DeleteTableRequest\032\024.DeleteTableResp" + + "onse\0228\n\013EnableTable\022\023.EnableTableRequest" + + "\032\024.EnableTableResponse\022;\n\014DisableTable\022\024" + + ".DisableTableRequest\032\025.DisableTableRespo", + "nse\0228\n\013ModifyTable\022\023.ModifyTableRequest\032" + + "\024.ModifyTableResponse\0228\n\013CreateTable\022\023.C" + + "reateTableRequest\032\024.CreateTableResponse\022" + + "/\n\010Shutdown\022\020.ShutdownRequest\032\021.Shutdown" + + "Response\0225\n\nStopMaster\022\022.StopMasterReque" + + "st\032\023.StopMasterResponse\022,\n\007Balance\022\017.Bal" + + "anceRequest\032\020.BalanceResponse\022M\n\022SetBala" + + "ncerRunning\022\032.SetBalancerRunningRequest\032" + + "\033.SetBalancerRunningResponse\022;\n\016RunCatal" + + "ogScan\022\023.CatalogScanRequest\032\024.CatalogSca", + "nResponse\022S\n\024EnableCatalogJanitor\022\034.Enab" + + "leCatalogJanitorRequest\032\035.EnableCatalogJ" + + "anitorResponse\022\\\n\027IsCatalogJanitorEnable" + + "d\022\037.IsCatalogJanitorEnabledRequest\032 .IsC" + + "atalogJanitorEnabledResponse\022L\n\021ExecMast" + + "erService\022\032.CoprocessorServiceRequest\032\033." + + "CoprocessorServiceResponse\0227\n\010Snapshot\022\024" + + ".TakeSnapshotRequest\032\025.TakeSnapshotRespo" + + "nse\022D\n\025GetCompletedSnapshots\022\024.ListSnaps" + + "hotRequest\032\025.ListSnapshotResponse\022A\n\016Del", + "eteSnapshot\022\026.DeleteSnapshotRequest\032\027.De" + + "leteSnapshotResponse\022A\n\016IsSnapshotDone\022\026" + + ".IsSnapshotDoneRequest\032\027.IsSnapshotDoneR" + + "esponse\022D\n\017RestoreSnapshot\022\027.RestoreSnap" + + "shotRequest\032\030.RestoreSnapshotResponse\022V\n" + + "\025IsRestoreSnapshotDone\022\035.IsRestoreSnapsh" + + "otDoneRequest\032\036.IsRestoreSnapshotDoneRes" + + "ponse\022D\n\017IsMasterRunning\022\027.IsMasterRunni" + + "ngRequest\032\030.IsMasterRunningResponseBG\n*o" + + "rg.apache.hadoop.hbase.protobuf.generate", + "dB\021MasterAdminProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java index c97ef67..af2b975 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java @@ -11,7 +11,7 @@ public final class MasterMonitorProtos { public interface GetSchemaAlterStatusRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); } @@ -44,8 +44,8 @@ public final class MasterMonitorProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -127,7 +127,7 @@ public final class MasterMonitorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -355,7 +355,7 @@ public final class MasterMonitorProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -393,11 +393,11 @@ public final class MasterMonitorProtos { public interface GetSchemaAlterStatusResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint32 yetToUpdateRegions = 1; + // optional uint32 yet_to_update_regions = 1; boolean hasYetToUpdateRegions(); int getYetToUpdateRegions(); - // optional uint32 totalRegions = 2; + // optional uint32 total_regions = 2; boolean hasTotalRegions(); int getTotalRegions(); } @@ -430,8 +430,8 @@ public final class MasterMonitorProtos { } private int bitField0_; - // optional uint32 yetToUpdateRegions = 1; - public static final int YETTOUPDATEREGIONS_FIELD_NUMBER = 1; + // optional uint32 yet_to_update_regions = 1; + public static final int YET_TO_UPDATE_REGIONS_FIELD_NUMBER = 1; private int yetToUpdateRegions_; public boolean hasYetToUpdateRegions() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -440,8 +440,8 @@ public final class MasterMonitorProtos { return yetToUpdateRegions_; } - // optional uint32 totalRegions = 2; - public static final int TOTALREGIONS_FIELD_NUMBER = 2; + // optional uint32 total_regions = 2; + public static final int TOTAL_REGIONS_FIELD_NUMBER = 2; private int totalRegions_; public boolean hasTotalRegions() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -532,11 +532,11 @@ public final class MasterMonitorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasYetToUpdateRegions()) { - hash = (37 * hash) + YETTOUPDATEREGIONS_FIELD_NUMBER; + hash = (37 * hash) + YET_TO_UPDATE_REGIONS_FIELD_NUMBER; hash = (53 * hash) + getYetToUpdateRegions(); } if (hasTotalRegions()) { - hash = (37 * hash) + TOTALREGIONS_FIELD_NUMBER; + hash = (37 * hash) + TOTAL_REGIONS_FIELD_NUMBER; hash = (53 * hash) + getTotalRegions(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -774,7 +774,7 @@ public final class MasterMonitorProtos { private int bitField0_; - // optional uint32 yetToUpdateRegions = 1; + // optional uint32 yet_to_update_regions = 1; private int yetToUpdateRegions_ ; public boolean hasYetToUpdateRegions() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -795,7 +795,7 @@ public final class MasterMonitorProtos { return this; } - // optional uint32 totalRegions = 2; + // optional uint32 total_regions = 2; private int totalRegions_ ; public boolean hasTotalRegions() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -830,7 +830,7 @@ public final class MasterMonitorProtos { public interface GetTableDescriptorsRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated string tableNames = 1; + // repeated string table_names = 1; java.util.List getTableNamesList(); int getTableNamesCount(); String getTableNames(int index); @@ -863,8 +863,8 @@ public final class MasterMonitorProtos { return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable; } - // repeated string tableNames = 1; - public static final int TABLENAMES_FIELD_NUMBER = 1; + // repeated string table_names = 1; + public static final int TABLE_NAMES_FIELD_NUMBER = 1; private com.google.protobuf.LazyStringList tableNames_; public java.util.List getTableNamesList() { @@ -948,7 +948,7 @@ public final class MasterMonitorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getTableNamesCount() > 0) { - hash = (37 * hash) + TABLENAMES_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER; hash = (53 * hash) + getTableNamesList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1179,7 +1179,7 @@ public final class MasterMonitorProtos { private int bitField0_; - // repeated string tableNames = 1; + // repeated string table_names = 1; private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureTableNamesIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { @@ -1249,7 +1249,7 @@ public final class MasterMonitorProtos { public interface GetTableDescriptorsResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .TableSchema tableSchema = 1; + // repeated .TableSchema table_schema = 1; java.util.List getTableSchemaList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index); @@ -1287,8 +1287,8 @@ public final class MasterMonitorProtos { return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable; } - // repeated .TableSchema tableSchema = 1; - public static final int TABLESCHEMA_FIELD_NUMBER = 1; + // repeated .TableSchema table_schema = 1; + public static final int TABLE_SCHEMA_FIELD_NUMBER = 1; private java.util.List tableSchema_; public java.util.List getTableSchemaList() { return tableSchema_; @@ -1380,7 +1380,7 @@ public final class MasterMonitorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getTableSchemaCount() > 0) { - hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; + hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchemaList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1642,7 +1642,7 @@ public final class MasterMonitorProtos { private int bitField0_; - // repeated .TableSchema tableSchema = 1; + // repeated .TableSchema table_schema = 1; private java.util.List tableSchema_ = java.util.Collections.emptyList(); private void ensureTableSchemaIsMutable() { @@ -2142,7 +2142,7 @@ public final class MasterMonitorProtos { public interface GetClusterStatusResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .ClusterStatus clusterStatus = 1; + // required .ClusterStatus cluster_status = 1; boolean hasClusterStatus(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder(); @@ -2176,8 +2176,8 @@ public final class MasterMonitorProtos { } private int bitField0_; - // required .ClusterStatus clusterStatus = 1; - public static final int CLUSTERSTATUS_FIELD_NUMBER = 1; + // required .ClusterStatus cluster_status = 1; + public static final int CLUSTER_STATUS_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_; public boolean hasClusterStatus() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2266,7 +2266,7 @@ public final class MasterMonitorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasClusterStatus()) { - hash = (37 * hash) + CLUSTERSTATUS_FIELD_NUMBER; + hash = (37 * hash) + CLUSTER_STATUS_FIELD_NUMBER; hash = (53 * hash) + getClusterStatus().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2511,7 +2511,7 @@ public final class MasterMonitorProtos { private int bitField0_; - // required .ClusterStatus clusterStatus = 1; + // required .ClusterStatus cluster_status = 1; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_; @@ -3068,26 +3068,26 @@ public final class MasterMonitorProtos { static { java.lang.String[] descriptorData = { "\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" + - "se.proto\032\023ClusterStatus.proto\"0\n\033GetSche" + - "maAlterStatusRequest\022\021\n\ttableName\030\001 \002(\014\"" + - "P\n\034GetSchemaAlterStatusResponse\022\032\n\022yetTo" + - "UpdateRegions\030\001 \001(\r\022\024\n\014totalRegions\030\002 \001(" + - "\r\"0\n\032GetTableDescriptorsRequest\022\022\n\ntable" + - "Names\030\001 \003(\t\"@\n\033GetTableDescriptorsRespon" + - "se\022!\n\013tableSchema\030\001 \003(\0132\014.TableSchema\"\031\n" + - "\027GetClusterStatusRequest\"A\n\030GetClusterSt" + - "atusResponse\022%\n\rclusterStatus\030\001 \002(\0132\016.Cl", - "usterStatus2\314\002\n\024MasterMonitorService\022S\n\024" + - "getSchemaAlterStatus\022\034.GetSchemaAlterSta" + - "tusRequest\032\035.GetSchemaAlterStatusRespons" + - "e\022P\n\023getTableDescriptors\022\033.GetTableDescr" + - "iptorsRequest\032\034.GetTableDescriptorsRespo" + - "nse\022G\n\020getClusterStatus\022\030.GetClusterStat" + - "usRequest\032\031.GetClusterStatusResponse\022D\n\017" + - "isMasterRunning\022\027.IsMasterRunningRequest" + - "\032\030.IsMasterRunningResponseBI\n*org.apache" + - ".hadoop.hbase.protobuf.generatedB\023Master", - "MonitorProtosH\001\210\001\001\240\001\001" + "se.proto\032\023ClusterStatus.proto\"1\n\033GetSche" + + "maAlterStatusRequest\022\022\n\ntable_name\030\001 \002(\014" + + "\"T\n\034GetSchemaAlterStatusResponse\022\035\n\025yet_" + + "to_update_regions\030\001 \001(\r\022\025\n\rtotal_regions" + + "\030\002 \001(\r\"1\n\032GetTableDescriptorsRequest\022\023\n\013" + + "table_names\030\001 \003(\t\"A\n\033GetTableDescriptors" + + "Response\022\"\n\014table_schema\030\001 \003(\0132\014.TableSc" + + "hema\"\031\n\027GetClusterStatusRequest\"B\n\030GetCl" + + "usterStatusResponse\022&\n\016cluster_status\030\001 ", + "\002(\0132\016.ClusterStatus2\314\002\n\024MasterMonitorSer" + + "vice\022S\n\024GetSchemaAlterStatus\022\034.GetSchema" + + "AlterStatusRequest\032\035.GetSchemaAlterStatu" + + "sResponse\022P\n\023GetTableDescriptors\022\033.GetTa" + + "bleDescriptorsRequest\032\034.GetTableDescript" + + "orsResponse\022G\n\020GetClusterStatus\022\030.GetClu" + + "sterStatusRequest\032\031.GetClusterStatusResp" + + "onse\022D\n\017IsMasterRunning\022\027.IsMasterRunnin" + + "gRequest\032\030.IsMasterRunningResponseBI\n*or" + + "g.apache.hadoop.hbase.protobuf.generated", + "B\023MasterMonitorProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java index b8c322f..ecfe7d2 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java @@ -311,7 +311,7 @@ public final class MasterProtos { public interface IsMasterRunningResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bool isMasterRunning = 1; + // required bool is_master_running = 1; boolean hasIsMasterRunning(); boolean getIsMasterRunning(); } @@ -344,8 +344,8 @@ public final class MasterProtos { } private int bitField0_; - // required bool isMasterRunning = 1; - public static final int ISMASTERRUNNING_FIELD_NUMBER = 1; + // required bool is_master_running = 1; + public static final int IS_MASTER_RUNNING_FIELD_NUMBER = 1; private boolean isMasterRunning_; public boolean hasIsMasterRunning() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -427,7 +427,7 @@ public final class MasterProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasIsMasterRunning()) { - hash = (37 * hash) + ISMASTERRUNNING_FIELD_NUMBER; + hash = (37 * hash) + IS_MASTER_RUNNING_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getIsMasterRunning()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -655,7 +655,7 @@ public final class MasterProtos { private int bitField0_; - // required bool isMasterRunning = 1; + // required bool is_master_running = 1; private boolean isMasterRunning_ ; public boolean hasIsMasterRunning() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -928,12 +928,12 @@ public final class MasterProtos { static { java.lang.String[] descriptorData = { "\n\014Master.proto\"\030\n\026IsMasterRunningRequest" + - "\"2\n\027IsMasterRunningResponse\022\027\n\017isMasterR" + - "unning\030\001 \002(\0102U\n\rMasterService\022D\n\017isMaste" + - "rRunning\022\027.IsMasterRunningRequest\032\030.IsMa" + - "sterRunningResponseBB\n*org.apache.hadoop" + - ".hbase.protobuf.generatedB\014MasterProtosH" + - "\001\210\001\001\240\001\001" + "\"4\n\027IsMasterRunningResponse\022\031\n\021is_master" + + "_running\030\001 \002(\0102U\n\rMasterService\022D\n\017IsMas" + + "terRunning\022\027.IsMasterRunningRequest\032\030.Is" + + "MasterRunningResponseBB\n*org.apache.hado" + + "op.hbase.protobuf.generatedB\014MasterProto" + + "sH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MultiRowMutation.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MultiRowMutation.java index 3227635..1eaeeb3 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MultiRowMutation.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MultiRowMutation.java @@ -11,7 +11,7 @@ public final class MultiRowMutation { public interface MultiMutateRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .MutationProto mutationRequest = 1; + // repeated .MutationProto mutation_request = 1; java.util.List getMutationRequestList(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index); @@ -49,8 +49,8 @@ public final class MultiRowMutation { return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable; } - // repeated .MutationProto mutationRequest = 1; - public static final int MUTATIONREQUEST_FIELD_NUMBER = 1; + // repeated .MutationProto mutation_request = 1; + public static final int MUTATION_REQUEST_FIELD_NUMBER = 1; private java.util.List mutationRequest_; public java.util.List getMutationRequestList() { return mutationRequest_; @@ -142,7 +142,7 @@ public final class MultiRowMutation { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getMutationRequestCount() > 0) { - hash = (37 * hash) + MUTATIONREQUEST_FIELD_NUMBER; + hash = (37 * hash) + MUTATION_REQUEST_FIELD_NUMBER; hash = (53 * hash) + getMutationRequestList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -404,7 +404,7 @@ public final class MultiRowMutation { private int bitField0_; - // repeated .MutationProto mutationRequest = 1; + // repeated .MutationProto mutation_request = 1; private java.util.List mutationRequest_ = java.util.Collections.emptyList(); private void ensureMutationRequestIsMutable() { @@ -1141,13 +1141,14 @@ public final class MultiRowMutation { descriptor; static { java.lang.String[] descriptorData = { - "\n\026MultiRowMutation.proto\032\014Client.proto\"=" + - "\n\022MultiMutateRequest\022\'\n\017mutationRequest\030" + - "\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateResp" + - "onse2R\n\027MultiRowMutationService\0227\n\nmutat" + - "eRows\022\023.MultiMutateRequest\032\024.MultiMutate" + - "ResponseBF\n*org.apache.hadoop.hbase.prot" + - "obuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001\001" + "\n\026MultiRowMutation.proto\032\014Client.proto\">" + + "\n\022MultiMutateRequest\022(\n\020mutation_request" + + "\030\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateRes" + + "ponse2R\n\027MultiRowMutationService\0227\n\nMuta" + + "teRows\022\023.MultiMutateRequest\032\024.MultiMutat" + + "eResponseBF\n*org.apache.hadoop.hbase.pro" + + "tobuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001" + + "\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java index 4cdaffc..2169655 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java @@ -11,11 +11,11 @@ public final class RPCProtos { public interface UserInformationOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string effectiveUser = 1; + // required string effective_user = 1; boolean hasEffectiveUser(); String getEffectiveUser(); - // optional string realUser = 2; + // optional string real_user = 2; boolean hasRealUser(); String getRealUser(); } @@ -48,8 +48,8 @@ public final class RPCProtos { } private int bitField0_; - // required string effectiveUser = 1; - public static final int EFFECTIVEUSER_FIELD_NUMBER = 1; + // required string effective_user = 1; + public static final int EFFECTIVE_USER_FIELD_NUMBER = 1; private java.lang.Object effectiveUser_; public boolean hasEffectiveUser() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -80,8 +80,8 @@ public final class RPCProtos { } } - // optional string realUser = 2; - public static final int REALUSER_FIELD_NUMBER = 2; + // optional string real_user = 2; + public static final int REAL_USER_FIELD_NUMBER = 2; private java.lang.Object realUser_; public boolean hasRealUser() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -198,11 +198,11 @@ public final class RPCProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasEffectiveUser()) { - hash = (37 * hash) + EFFECTIVEUSER_FIELD_NUMBER; + hash = (37 * hash) + EFFECTIVE_USER_FIELD_NUMBER; hash = (53 * hash) + getEffectiveUser().hashCode(); } if (hasRealUser()) { - hash = (37 * hash) + REALUSER_FIELD_NUMBER; + hash = (37 * hash) + REAL_USER_FIELD_NUMBER; hash = (53 * hash) + getRealUser().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -444,7 +444,7 @@ public final class RPCProtos { private int bitField0_; - // required string effectiveUser = 1; + // required string effective_user = 1; private java.lang.Object effectiveUser_ = ""; public boolean hasEffectiveUser() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -480,7 +480,7 @@ public final class RPCProtos { onChanged(); } - // optional string realUser = 2; + // optional string real_user = 2; private java.lang.Object realUser_ = ""; public boolean hasRealUser() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -530,20 +530,20 @@ public final class RPCProtos { public interface ConnectionHeaderOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional .UserInformation userInfo = 1; + // optional .UserInformation user_info = 1; boolean hasUserInfo(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder(); - // optional string serviceName = 2; + // optional string service_name = 2; boolean hasServiceName(); String getServiceName(); - // optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; + // optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; boolean hasCellBlockCodecClass(); String getCellBlockCodecClass(); - // optional string cellBlockCompressorClass = 4; + // optional string cell_block_compressor_class = 4; boolean hasCellBlockCompressorClass(); String getCellBlockCompressorClass(); } @@ -576,8 +576,8 @@ public final class RPCProtos { } private int bitField0_; - // optional .UserInformation userInfo = 1; - public static final int USERINFO_FIELD_NUMBER = 1; + // optional .UserInformation user_info = 1; + public static final int USER_INFO_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_; public boolean hasUserInfo() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -589,8 +589,8 @@ public final class RPCProtos { return userInfo_; } - // optional string serviceName = 2; - public static final int SERVICENAME_FIELD_NUMBER = 2; + // optional string service_name = 2; + public static final int SERVICE_NAME_FIELD_NUMBER = 2; private java.lang.Object serviceName_; public boolean hasServiceName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -621,8 +621,8 @@ public final class RPCProtos { } } - // optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; - public static final int CELLBLOCKCODECCLASS_FIELD_NUMBER = 3; + // optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; + public static final int CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER = 3; private java.lang.Object cellBlockCodecClass_; public boolean hasCellBlockCodecClass() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -653,8 +653,8 @@ public final class RPCProtos { } } - // optional string cellBlockCompressorClass = 4; - public static final int CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER = 4; + // optional string cell_block_compressor_class = 4; + public static final int CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER = 4; private java.lang.Object cellBlockCompressorClass_; public boolean hasCellBlockCompressorClass() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -799,19 +799,19 @@ public final class RPCProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasUserInfo()) { - hash = (37 * hash) + USERINFO_FIELD_NUMBER; + hash = (37 * hash) + USER_INFO_FIELD_NUMBER; hash = (53 * hash) + getUserInfo().hashCode(); } if (hasServiceName()) { - hash = (37 * hash) + SERVICENAME_FIELD_NUMBER; + hash = (37 * hash) + SERVICE_NAME_FIELD_NUMBER; hash = (53 * hash) + getServiceName().hashCode(); } if (hasCellBlockCodecClass()) { - hash = (37 * hash) + CELLBLOCKCODECCLASS_FIELD_NUMBER; + hash = (37 * hash) + CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER; hash = (53 * hash) + getCellBlockCodecClass().hashCode(); } if (hasCellBlockCompressorClass()) { - hash = (37 * hash) + CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER; + hash = (37 * hash) + CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER; hash = (53 * hash) + getCellBlockCompressorClass().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1096,7 +1096,7 @@ public final class RPCProtos { private int bitField0_; - // optional .UserInformation userInfo = 1; + // optional .UserInformation user_info = 1; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_; @@ -1186,7 +1186,7 @@ public final class RPCProtos { return userInfoBuilder_; } - // optional string serviceName = 2; + // optional string service_name = 2; private java.lang.Object serviceName_ = ""; public boolean hasServiceName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1222,7 +1222,7 @@ public final class RPCProtos { onChanged(); } - // optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; + // optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec"; public boolean hasCellBlockCodecClass() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1258,7 +1258,7 @@ public final class RPCProtos { onChanged(); } - // optional string cellBlockCompressorClass = 4; + // optional string cell_block_compressor_class = 4; private java.lang.Object cellBlockCompressorClass_ = ""; public boolean hasCellBlockCompressorClass() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1679,11 +1679,11 @@ public final class RPCProtos { public interface ExceptionResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional string exceptionClassName = 1; + // optional string exception_class_name = 1; boolean hasExceptionClassName(); String getExceptionClassName(); - // optional string stackTrace = 2; + // optional string stack_trace = 2; boolean hasStackTrace(); String getStackTrace(); @@ -1695,7 +1695,7 @@ public final class RPCProtos { boolean hasPort(); int getPort(); - // optional bool doNotRetry = 5; + // optional bool do_not_retry = 5; boolean hasDoNotRetry(); boolean getDoNotRetry(); } @@ -1728,8 +1728,8 @@ public final class RPCProtos { } private int bitField0_; - // optional string exceptionClassName = 1; - public static final int EXCEPTIONCLASSNAME_FIELD_NUMBER = 1; + // optional string exception_class_name = 1; + public static final int EXCEPTION_CLASS_NAME_FIELD_NUMBER = 1; private java.lang.Object exceptionClassName_; public boolean hasExceptionClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1760,8 +1760,8 @@ public final class RPCProtos { } } - // optional string stackTrace = 2; - public static final int STACKTRACE_FIELD_NUMBER = 2; + // optional string stack_trace = 2; + public static final int STACK_TRACE_FIELD_NUMBER = 2; private java.lang.Object stackTrace_; public boolean hasStackTrace() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1834,8 +1834,8 @@ public final class RPCProtos { return port_; } - // optional bool doNotRetry = 5; - public static final int DONOTRETRY_FIELD_NUMBER = 5; + // optional bool do_not_retry = 5; + public static final int DO_NOT_RETRY_FIELD_NUMBER = 5; private boolean doNotRetry_; public boolean hasDoNotRetry() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -1965,11 +1965,11 @@ public final class RPCProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasExceptionClassName()) { - hash = (37 * hash) + EXCEPTIONCLASSNAME_FIELD_NUMBER; + hash = (37 * hash) + EXCEPTION_CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getExceptionClassName().hashCode(); } if (hasStackTrace()) { - hash = (37 * hash) + STACKTRACE_FIELD_NUMBER; + hash = (37 * hash) + STACK_TRACE_FIELD_NUMBER; hash = (53 * hash) + getStackTrace().hashCode(); } if (hasHostname()) { @@ -1981,7 +1981,7 @@ public final class RPCProtos { hash = (53 * hash) + getPort(); } if (hasDoNotRetry()) { - hash = (37 * hash) + DONOTRETRY_FIELD_NUMBER; + hash = (37 * hash) + DO_NOT_RETRY_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getDoNotRetry()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2261,7 +2261,7 @@ public final class RPCProtos { private int bitField0_; - // optional string exceptionClassName = 1; + // optional string exception_class_name = 1; private java.lang.Object exceptionClassName_ = ""; public boolean hasExceptionClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2297,7 +2297,7 @@ public final class RPCProtos { onChanged(); } - // optional string stackTrace = 2; + // optional string stack_trace = 2; private java.lang.Object stackTrace_ = ""; public boolean hasStackTrace() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2390,7 +2390,7 @@ public final class RPCProtos { return this; } - // optional bool doNotRetry = 5; + // optional bool do_not_retry = 5; private boolean doNotRetry_ ; public boolean hasDoNotRetry() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -2425,24 +2425,24 @@ public final class RPCProtos { public interface RequestHeaderOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint32 callId = 1; + // optional uint32 call_id = 1; boolean hasCallId(); int getCallId(); - // optional .RPCTInfo traceInfo = 2; + // optional .RPCTInfo trace_info = 2; boolean hasTraceInfo(); org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo(); org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder(); - // optional string methodName = 3; + // optional string method_name = 3; boolean hasMethodName(); String getMethodName(); - // optional bool requestParam = 4; + // optional bool request_param = 4; boolean hasRequestParam(); boolean getRequestParam(); - // optional .CellBlockMeta cellBlockMeta = 5; + // optional .CellBlockMeta cell_block_meta = 5; boolean hasCellBlockMeta(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder(); @@ -2476,8 +2476,8 @@ public final class RPCProtos { } private int bitField0_; - // optional uint32 callId = 1; - public static final int CALLID_FIELD_NUMBER = 1; + // optional uint32 call_id = 1; + public static final int CALL_ID_FIELD_NUMBER = 1; private int callId_; public boolean hasCallId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2486,8 +2486,8 @@ public final class RPCProtos { return callId_; } - // optional .RPCTInfo traceInfo = 2; - public static final int TRACEINFO_FIELD_NUMBER = 2; + // optional .RPCTInfo trace_info = 2; + public static final int TRACE_INFO_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_; public boolean hasTraceInfo() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2499,8 +2499,8 @@ public final class RPCProtos { return traceInfo_; } - // optional string methodName = 3; - public static final int METHODNAME_FIELD_NUMBER = 3; + // optional string method_name = 3; + public static final int METHOD_NAME_FIELD_NUMBER = 3; private java.lang.Object methodName_; public boolean hasMethodName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2531,8 +2531,8 @@ public final class RPCProtos { } } - // optional bool requestParam = 4; - public static final int REQUESTPARAM_FIELD_NUMBER = 4; + // optional bool request_param = 4; + public static final int REQUEST_PARAM_FIELD_NUMBER = 4; private boolean requestParam_; public boolean hasRequestParam() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2541,8 +2541,8 @@ public final class RPCProtos { return requestParam_; } - // optional .CellBlockMeta cellBlockMeta = 5; - public static final int CELLBLOCKMETA_FIELD_NUMBER = 5; + // optional .CellBlockMeta cell_block_meta = 5; + public static final int CELL_BLOCK_META_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_; public boolean hasCellBlockMeta() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -2675,23 +2675,23 @@ public final class RPCProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCallId()) { - hash = (37 * hash) + CALLID_FIELD_NUMBER; + hash = (37 * hash) + CALL_ID_FIELD_NUMBER; hash = (53 * hash) + getCallId(); } if (hasTraceInfo()) { - hash = (37 * hash) + TRACEINFO_FIELD_NUMBER; + hash = (37 * hash) + TRACE_INFO_FIELD_NUMBER; hash = (53 * hash) + getTraceInfo().hashCode(); } if (hasMethodName()) { - hash = (37 * hash) + METHODNAME_FIELD_NUMBER; + hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER; hash = (53 * hash) + getMethodName().hashCode(); } if (hasRequestParam()) { - hash = (37 * hash) + REQUESTPARAM_FIELD_NUMBER; + hash = (37 * hash) + REQUEST_PARAM_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getRequestParam()); } if (hasCellBlockMeta()) { - hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER; + hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER; hash = (53 * hash) + getCellBlockMeta().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2997,7 +2997,7 @@ public final class RPCProtos { private int bitField0_; - // optional uint32 callId = 1; + // optional uint32 call_id = 1; private int callId_ ; public boolean hasCallId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3018,7 +3018,7 @@ public final class RPCProtos { return this; } - // optional .RPCTInfo traceInfo = 2; + // optional .RPCTInfo trace_info = 2; private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_; @@ -3108,7 +3108,7 @@ public final class RPCProtos { return traceInfoBuilder_; } - // optional string methodName = 3; + // optional string method_name = 3; private java.lang.Object methodName_ = ""; public boolean hasMethodName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -3144,7 +3144,7 @@ public final class RPCProtos { onChanged(); } - // optional bool requestParam = 4; + // optional bool request_param = 4; private boolean requestParam_ ; public boolean hasRequestParam() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -3165,7 +3165,7 @@ public final class RPCProtos { return this; } - // optional .CellBlockMeta cellBlockMeta = 5; + // optional .CellBlockMeta cell_block_meta = 5; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_; @@ -3269,7 +3269,7 @@ public final class RPCProtos { public interface ResponseHeaderOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional uint32 callId = 1; + // optional uint32 call_id = 1; boolean hasCallId(); int getCallId(); @@ -3278,7 +3278,7 @@ public final class RPCProtos { org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder(); - // optional .CellBlockMeta cellBlockMeta = 3; + // optional .CellBlockMeta cell_block_meta = 3; boolean hasCellBlockMeta(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder(); @@ -3312,8 +3312,8 @@ public final class RPCProtos { } private int bitField0_; - // optional uint32 callId = 1; - public static final int CALLID_FIELD_NUMBER = 1; + // optional uint32 call_id = 1; + public static final int CALL_ID_FIELD_NUMBER = 1; private int callId_; public boolean hasCallId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3335,8 +3335,8 @@ public final class RPCProtos { return exception_; } - // optional .CellBlockMeta cellBlockMeta = 3; - public static final int CELLBLOCKMETA_FIELD_NUMBER = 3; + // optional .CellBlockMeta cell_block_meta = 3; + public static final int CELL_BLOCK_META_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_; public boolean hasCellBlockMeta() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -3443,7 +3443,7 @@ public final class RPCProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasCallId()) { - hash = (37 * hash) + CALLID_FIELD_NUMBER; + hash = (37 * hash) + CALL_ID_FIELD_NUMBER; hash = (53 * hash) + getCallId(); } if (hasException()) { @@ -3451,7 +3451,7 @@ public final class RPCProtos { hash = (53 * hash) + getException().hashCode(); } if (hasCellBlockMeta()) { - hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER; + hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER; hash = (53 * hash) + getCellBlockMeta().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3729,7 +3729,7 @@ public final class RPCProtos { private int bitField0_; - // optional uint32 callId = 1; + // optional uint32 call_id = 1; private int callId_ ; public boolean hasCallId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3840,7 +3840,7 @@ public final class RPCProtos { return exceptionBuilder_; } - // optional .CellBlockMeta cellBlockMeta = 3; + // optional .CellBlockMeta cell_block_meta = 3; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_; @@ -3981,24 +3981,25 @@ public final class RPCProtos { static { java.lang.String[] descriptorData = { "\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" + - ":\n\017UserInformation\022\025\n\reffectiveUser\030\001 \002(" + - "\t\022\020\n\010realUser\030\002 \001(\t\"\267\001\n\020ConnectionHeader" + - "\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022\023\n\013" + - "serviceName\030\002 \001(\t\022H\n\023cellBlockCodecClass" + - "\030\003 \001(\t:+org.apache.hadoop.hbase.codec.Ke" + - "yValueCodec\022 \n\030cellBlockCompressorClass\030" + - "\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001(\r\"w" + - "\n\021ExceptionResponse\022\032\n\022exceptionClassNam" + - "e\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostname\030", - "\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 \001(\010\"" + - "\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n\ttra" + - "ceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodName\030\003 " + - "\001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlockMe" + - "ta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016ResponseHead" + - "er\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(\0132\022." + - "ExceptionResponse\022%\n\rcellBlockMeta\030\003 \001(\013" + - "2\016.CellBlockMetaB<\n*org.apache.hadoop.hb" + - "ase.protobuf.generatedB\tRPCProtosH\001\240\001\001" + "<\n\017UserInformation\022\026\n\016effective_user\030\001 \002" + + "(\t\022\021\n\treal_user\030\002 \001(\t\"\277\001\n\020ConnectionHead" + + "er\022#\n\tuser_info\030\001 \001(\0132\020.UserInformation\022" + + "\024\n\014service_name\030\002 \001(\t\022K\n\026cell_block_code" + + "c_class\030\003 \001(\t:+org.apache.hadoop.hbase.c" + + "odec.KeyValueCodec\022#\n\033cell_block_compres" + + "sor_class\030\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006leng" + + "th\030\001 \001(\r\"|\n\021ExceptionResponse\022\034\n\024excepti" + + "on_class_name\030\001 \001(\t\022\023\n\013stack_trace\030\002 \001(\t", + "\022\020\n\010hostname\030\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\024\n\014do_n" + + "ot_retry\030\005 \001(\010\"\224\001\n\rRequestHeader\022\017\n\007call" + + "_id\030\001 \001(\r\022\035\n\ntrace_info\030\002 \001(\0132\t.RPCTInfo" + + "\022\023\n\013method_name\030\003 \001(\t\022\025\n\rrequest_param\030\004" + + " \001(\010\022\'\n\017cell_block_meta\030\005 \001(\0132\016.CellBloc" + + "kMeta\"q\n\016ResponseHeader\022\017\n\007call_id\030\001 \001(\r" + + "\022%\n\texception\030\002 \001(\0132\022.ExceptionResponse\022" + + "\'\n\017cell_block_meta\030\003 \001(\0132\016.CellBlockMeta" + + "B<\n*org.apache.hadoop.hbase.protobuf.gen" + + "eratedB\tRPCProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java index c9f01a4..6da4a8a 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RegionServerStatusProtos.java @@ -15,11 +15,11 @@ public final class RegionServerStatusProtos { boolean hasPort(); int getPort(); - // required uint64 serverStartCode = 2; + // required uint64 server_start_code = 2; boolean hasServerStartCode(); long getServerStartCode(); - // required uint64 serverCurrentTime = 3; + // required uint64 server_current_time = 3; boolean hasServerCurrentTime(); long getServerCurrentTime(); } @@ -62,8 +62,8 @@ public final class RegionServerStatusProtos { return port_; } - // required uint64 serverStartCode = 2; - public static final int SERVERSTARTCODE_FIELD_NUMBER = 2; + // required uint64 server_start_code = 2; + public static final int SERVER_START_CODE_FIELD_NUMBER = 2; private long serverStartCode_; public boolean hasServerStartCode() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -72,8 +72,8 @@ public final class RegionServerStatusProtos { return serverStartCode_; } - // required uint64 serverCurrentTime = 3; - public static final int SERVERCURRENTTIME_FIELD_NUMBER = 3; + // required uint64 server_current_time = 3; + public static final int SERVER_CURRENT_TIME_FIELD_NUMBER = 3; private long serverCurrentTime_; public boolean hasServerCurrentTime() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -193,11 +193,11 @@ public final class RegionServerStatusProtos { hash = (53 * hash) + getPort(); } if (hasServerStartCode()) { - hash = (37 * hash) + SERVERSTARTCODE_FIELD_NUMBER; + hash = (37 * hash) + SERVER_START_CODE_FIELD_NUMBER; hash = (53 * hash) + hashLong(getServerStartCode()); } if (hasServerCurrentTime()) { - hash = (37 * hash) + SERVERCURRENTTIME_FIELD_NUMBER; + hash = (37 * hash) + SERVER_CURRENT_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getServerCurrentTime()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -482,7 +482,7 @@ public final class RegionServerStatusProtos { return this; } - // required uint64 serverStartCode = 2; + // required uint64 server_start_code = 2; private long serverStartCode_ ; public boolean hasServerStartCode() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -503,7 +503,7 @@ public final class RegionServerStatusProtos { return this; } - // required uint64 serverCurrentTime = 3; + // required uint64 server_current_time = 3; private long serverCurrentTime_ ; public boolean hasServerCurrentTime() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -538,7 +538,7 @@ public final class RegionServerStatusProtos { public interface RegionServerStartupResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .NameStringPair mapEntries = 1; + // repeated .NameStringPair map_entries = 1; java.util.List getMapEntriesList(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index); @@ -576,8 +576,8 @@ public final class RegionServerStatusProtos { return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable; } - // repeated .NameStringPair mapEntries = 1; - public static final int MAPENTRIES_FIELD_NUMBER = 1; + // repeated .NameStringPair map_entries = 1; + public static final int MAP_ENTRIES_FIELD_NUMBER = 1; private java.util.List mapEntries_; public java.util.List getMapEntriesList() { return mapEntries_; @@ -669,7 +669,7 @@ public final class RegionServerStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getMapEntriesCount() > 0) { - hash = (37 * hash) + MAPENTRIES_FIELD_NUMBER; + hash = (37 * hash) + MAP_ENTRIES_FIELD_NUMBER; hash = (53 * hash) + getMapEntriesList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -931,7 +931,7 @@ public final class RegionServerStatusProtos { private int bitField0_; - // repeated .NameStringPair mapEntries = 1; + // repeated .NameStringPair map_entries = 1; private java.util.List mapEntries_ = java.util.Collections.emptyList(); private void ensureMapEntriesIsMutable() { @@ -2073,7 +2073,7 @@ public final class RegionServerStatusProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); - // required string errorMessage = 2; + // required string error_message = 2; boolean hasErrorMessage(); String getErrorMessage(); } @@ -2119,8 +2119,8 @@ public final class RegionServerStatusProtos { return server_; } - // required string errorMessage = 2; - public static final int ERRORMESSAGE_FIELD_NUMBER = 2; + // required string error_message = 2; + public static final int ERROR_MESSAGE_FIELD_NUMBER = 2; private java.lang.Object errorMessage_; public boolean hasErrorMessage() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2249,7 +2249,7 @@ public final class RegionServerStatusProtos { hash = (53 * hash) + getServer().hashCode(); } if (hasErrorMessage()) { - hash = (37 * hash) + ERRORMESSAGE_FIELD_NUMBER; + hash = (37 * hash) + ERROR_MESSAGE_FIELD_NUMBER; hash = (53 * hash) + getErrorMessage().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2602,7 +2602,7 @@ public final class RegionServerStatusProtos { return serverBuilder_; } - // required string errorMessage = 2; + // required string error_message = 2; private java.lang.Object errorMessage_ = ""; public boolean hasErrorMessage() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2952,7 +2952,7 @@ public final class RegionServerStatusProtos { public interface GetLastFlushedSequenceIdRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes regionName = 1; + // required bytes region_name = 1; boolean hasRegionName(); com.google.protobuf.ByteString getRegionName(); } @@ -2985,8 +2985,8 @@ public final class RegionServerStatusProtos { } private int bitField0_; - // required bytes regionName = 1; - public static final int REGIONNAME_FIELD_NUMBER = 1; + // required bytes region_name = 1; + public static final int REGION_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString regionName_; public boolean hasRegionName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3068,7 +3068,7 @@ public final class RegionServerStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRegionName()) { - hash = (37 * hash) + REGIONNAME_FIELD_NUMBER; + hash = (37 * hash) + REGION_NAME_FIELD_NUMBER; hash = (53 * hash) + getRegionName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3296,7 +3296,7 @@ public final class RegionServerStatusProtos { private int bitField0_; - // required bytes regionName = 1; + // required bytes region_name = 1; private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasRegionName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3334,7 +3334,7 @@ public final class RegionServerStatusProtos { public interface GetLastFlushedSequenceIdResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required uint64 lastFlushedSequenceId = 1; + // required uint64 last_flushed_sequence_id = 1; boolean hasLastFlushedSequenceId(); long getLastFlushedSequenceId(); } @@ -3367,8 +3367,8 @@ public final class RegionServerStatusProtos { } private int bitField0_; - // required uint64 lastFlushedSequenceId = 1; - public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1; + // required uint64 last_flushed_sequence_id = 1; + public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1; private long lastFlushedSequenceId_; public boolean hasLastFlushedSequenceId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3450,7 +3450,7 @@ public final class RegionServerStatusProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLastFlushedSequenceId()) { - hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER; + hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLastFlushedSequenceId()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3678,7 +3678,7 @@ public final class RegionServerStatusProtos { private int bitField0_; - // required uint64 lastFlushedSequenceId = 1; + // required uint64 last_flushed_sequence_id = 1; private long lastFlushedSequenceId_ ; public boolean hasLastFlushedSequenceId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -4176,31 +4176,31 @@ public final class RegionServerStatusProtos { static { java.lang.String[] descriptorData = { "\n\030RegionServerStatus.proto\032\013hbase.proto\032" + - "\023ClusterStatus.proto\"^\n\032RegionServerStar" + - "tupRequest\022\014\n\004port\030\001 \002(\r\022\027\n\017serverStartC" + - "ode\030\002 \002(\004\022\031\n\021serverCurrentTime\030\003 \002(\004\"B\n\033" + - "RegionServerStartupResponse\022#\n\nmapEntrie" + - "s\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionServer" + - "ReportRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa" + - "me\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032RegionS" + - "erverReportResponse\"N\n\031ReportRSFatalErro" + - "rRequest\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\024\n", - "\014errorMessage\030\002 \002(\t\"\034\n\032ReportRSFatalErro" + - "rResponse\"5\n\037GetLastFlushedSequenceIdReq" + - "uest\022\022\n\nregionName\030\001 \002(\014\"A\n GetLastFlush" + - "edSequenceIdResponse\022\035\n\025lastFlushedSeque" + - "nceId\030\001 \002(\0042\354\002\n\031RegionServerStatusServic" + - "e\022P\n\023regionServerStartup\022\033.RegionServerS" + - "tartupRequest\032\034.RegionServerStartupRespo" + - "nse\022M\n\022regionServerReport\022\032.RegionServer" + - "ReportRequest\032\033.RegionServerReportRespon" + - "se\022M\n\022reportRSFatalError\022\032.ReportRSFatal", - "ErrorRequest\032\033.ReportRSFatalErrorRespons" + - "e\022_\n\030getLastFlushedSequenceId\022 .GetLastF" + - "lushedSequenceIdRequest\032!.GetLastFlushed" + - "SequenceIdResponseBN\n*org.apache.hadoop." + - "hbase.protobuf.generatedB\030RegionServerSt" + - "atusProtosH\001\210\001\001\240\001\001" + "\023ClusterStatus.proto\"b\n\032RegionServerStar" + + "tupRequest\022\014\n\004port\030\001 \002(\r\022\031\n\021server_start" + + "_code\030\002 \002(\004\022\033\n\023server_current_time\030\003 \002(\004" + + "\"C\n\033RegionServerStartupResponse\022$\n\013map_e" + + "ntries\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionS" + + "erverReportRequest\022\033\n\006server\030\001 \002(\0132\013.Ser" + + "verName\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032Re" + + "gionServerReportResponse\"O\n\031ReportRSFata" + + "lErrorRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa", + "me\022\025\n\rerror_message\030\002 \002(\t\"\034\n\032ReportRSFat" + + "alErrorResponse\"6\n\037GetLastFlushedSequenc" + + "eIdRequest\022\023\n\013region_name\030\001 \002(\014\"D\n GetLa" + + "stFlushedSequenceIdResponse\022 \n\030last_flus" + + "hed_sequence_id\030\001 \002(\0042\354\002\n\031RegionServerSt" + + "atusService\022P\n\023RegionServerStartup\022\033.Reg" + + "ionServerStartupRequest\032\034.RegionServerSt" + + "artupResponse\022M\n\022RegionServerReport\022\032.Re" + + "gionServerReportRequest\032\033.RegionServerRe" + + "portResponse\022M\n\022ReportRSFatalError\022\032.Rep", + "ortRSFatalErrorRequest\032\033.ReportRSFatalEr" + + "rorResponse\022_\n\030GetLastFlushedSequenceId\022" + + " .GetLastFlushedSequenceIdRequest\032!.GetL" + + "astFlushedSequenceIdResponseBN\n*org.apac" + + "he.hadoop.hbase.protobuf.generatedB\030Regi" + + "onServerStatusProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RowProcessorProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RowProcessorProtos.java index 162639e..cc6b4f4 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RowProcessorProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RowProcessorProtos.java @@ -11,15 +11,15 @@ public final class RowProcessorProtos { public interface RowProcessorRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string rowProcessorClassName = 1; + // required string row_processor_class_name = 1; boolean hasRowProcessorClassName(); String getRowProcessorClassName(); - // optional string rowProcessorInitializerMessageName = 2; + // optional string row_processor_initializer_message_name = 2; boolean hasRowProcessorInitializerMessageName(); String getRowProcessorInitializerMessageName(); - // optional bytes rowProcessorInitializerMessage = 3; + // optional bytes row_processor_initializer_message = 3; boolean hasRowProcessorInitializerMessage(); com.google.protobuf.ByteString getRowProcessorInitializerMessage(); } @@ -52,8 +52,8 @@ public final class RowProcessorProtos { } private int bitField0_; - // required string rowProcessorClassName = 1; - public static final int ROWPROCESSORCLASSNAME_FIELD_NUMBER = 1; + // required string row_processor_class_name = 1; + public static final int ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER = 1; private java.lang.Object rowProcessorClassName_; public boolean hasRowProcessorClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -84,8 +84,8 @@ public final class RowProcessorProtos { } } - // optional string rowProcessorInitializerMessageName = 2; - public static final int ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER = 2; + // optional string row_processor_initializer_message_name = 2; + public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER = 2; private java.lang.Object rowProcessorInitializerMessageName_; public boolean hasRowProcessorInitializerMessageName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -116,8 +116,8 @@ public final class RowProcessorProtos { } } - // optional bytes rowProcessorInitializerMessage = 3; - public static final int ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER = 3; + // optional bytes row_processor_initializer_message = 3; + public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER = 3; private com.google.protobuf.ByteString rowProcessorInitializerMessage_; public boolean hasRowProcessorInitializerMessage() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -225,15 +225,15 @@ public final class RowProcessorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRowProcessorClassName()) { - hash = (37 * hash) + ROWPROCESSORCLASSNAME_FIELD_NUMBER; + hash = (37 * hash) + ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getRowProcessorClassName().hashCode(); } if (hasRowProcessorInitializerMessageName()) { - hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER; + hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER; hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode(); } if (hasRowProcessorInitializerMessage()) { - hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER; + hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER; hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -489,7 +489,7 @@ public final class RowProcessorProtos { private int bitField0_; - // required string rowProcessorClassName = 1; + // required string row_processor_class_name = 1; private java.lang.Object rowProcessorClassName_ = ""; public boolean hasRowProcessorClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -525,7 +525,7 @@ public final class RowProcessorProtos { onChanged(); } - // optional string rowProcessorInitializerMessageName = 2; + // optional string row_processor_initializer_message_name = 2; private java.lang.Object rowProcessorInitializerMessageName_ = ""; public boolean hasRowProcessorInitializerMessageName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -561,7 +561,7 @@ public final class RowProcessorProtos { onChanged(); } - // optional bytes rowProcessorInitializerMessage = 3; + // optional bytes row_processor_initializer_message = 3; private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY; public boolean hasRowProcessorInitializerMessage() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -599,7 +599,7 @@ public final class RowProcessorProtos { public interface RowProcessorResultOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes rowProcessorResult = 1; + // required bytes row_processor_result = 1; boolean hasRowProcessorResult(); com.google.protobuf.ByteString getRowProcessorResult(); } @@ -632,8 +632,8 @@ public final class RowProcessorProtos { } private int bitField0_; - // required bytes rowProcessorResult = 1; - public static final int ROWPROCESSORRESULT_FIELD_NUMBER = 1; + // required bytes row_processor_result = 1; + public static final int ROW_PROCESSOR_RESULT_FIELD_NUMBER = 1; private com.google.protobuf.ByteString rowProcessorResult_; public boolean hasRowProcessorResult() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -715,7 +715,7 @@ public final class RowProcessorProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasRowProcessorResult()) { - hash = (37 * hash) + ROWPROCESSORRESULT_FIELD_NUMBER; + hash = (37 * hash) + ROW_PROCESSOR_RESULT_FIELD_NUMBER; hash = (53 * hash) + getRowProcessorResult().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -943,7 +943,7 @@ public final class RowProcessorProtos { private int bitField0_; - // required bytes rowProcessorResult = 1; + // required bytes row_processor_result = 1; private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY; public boolean hasRowProcessorResult() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1218,15 +1218,16 @@ public final class RowProcessorProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\022RowProcessor.proto\"\210\001\n\023RowProcessorReq" + - "uest\022\035\n\025rowProcessorClassName\030\001 \002(\t\022*\n\"r" + - "owProcessorInitializerMessageName\030\002 \001(\t\022" + - "&\n\036rowProcessorInitializerMessage\030\003 \001(\014\"" + - "0\n\022RowProcessorResult\022\032\n\022rowProcessorRes" + - "ult\030\001 \002(\0142K\n\023RowProcessorService\0224\n\007proc" + - "ess\022\024.RowProcessorRequest\032\023.RowProcessor" + - "ResultBH\n*org.apache.hadoop.hbase.protob" + - "uf.generatedB\022RowProcessorProtosH\001\210\001\001\240\001\001" + "\n\022RowProcessor.proto\"\222\001\n\023RowProcessorReq" + + "uest\022 \n\030row_processor_class_name\030\001 \002(\t\022." + + "\n&row_processor_initializer_message_name" + + "\030\002 \001(\t\022)\n!row_processor_initializer_mess" + + "age\030\003 \001(\014\"2\n\022RowProcessorResult\022\034\n\024row_p" + + "rocessor_result\030\001 \002(\0142K\n\023RowProcessorSer" + + "vice\0224\n\007Process\022\024.RowProcessorRequest\032\023." + + "RowProcessorResultBH\n*org.apache.hadoop." + + "hbase.protobuf.generatedB\022RowProcessorPr" + + "otosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java index 2766a7b..e39dc10 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java @@ -11,7 +11,7 @@ public final class SecureBulkLoadProtos { public interface SecureBulkLoadHFilesRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1; + // repeated .BulkLoadHFileRequest.FamilyPath family_path = 1; java.util.List getFamilyPathList(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index); @@ -21,16 +21,16 @@ public final class SecureBulkLoadProtos { org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder( int index); - // optional bool assignSeqNum = 2; + // optional bool assign_seq_num = 2; boolean hasAssignSeqNum(); boolean getAssignSeqNum(); - // required .DelegationTokenProto fsToken = 3; + // required .DelegationTokenProto fs_token = 3; boolean hasFsToken(); org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken(); org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder(); - // required string bulkToken = 4; + // required string bulk_token = 4; boolean hasBulkToken(); String getBulkToken(); } @@ -63,8 +63,8 @@ public final class SecureBulkLoadProtos { } private int bitField0_; - // repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1; - public static final int FAMILYPATH_FIELD_NUMBER = 1; + // repeated .BulkLoadHFileRequest.FamilyPath family_path = 1; + public static final int FAMILY_PATH_FIELD_NUMBER = 1; private java.util.List familyPath_; public java.util.List getFamilyPathList() { return familyPath_; @@ -84,8 +84,8 @@ public final class SecureBulkLoadProtos { return familyPath_.get(index); } - // optional bool assignSeqNum = 2; - public static final int ASSIGNSEQNUM_FIELD_NUMBER = 2; + // optional bool assign_seq_num = 2; + public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 2; private boolean assignSeqNum_; public boolean hasAssignSeqNum() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -94,8 +94,8 @@ public final class SecureBulkLoadProtos { return assignSeqNum_; } - // required .DelegationTokenProto fsToken = 3; - public static final int FSTOKEN_FIELD_NUMBER = 3; + // required .DelegationTokenProto fs_token = 3; + public static final int FS_TOKEN_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_; public boolean hasFsToken() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -107,8 +107,8 @@ public final class SecureBulkLoadProtos { return fsToken_; } - // required string bulkToken = 4; - public static final int BULKTOKEN_FIELD_NUMBER = 4; + // required string bulk_token = 4; + public static final int BULK_TOKEN_FIELD_NUMBER = 4; private java.lang.Object bulkToken_; public boolean hasBulkToken() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -258,19 +258,19 @@ public final class SecureBulkLoadProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (getFamilyPathCount() > 0) { - hash = (37 * hash) + FAMILYPATH_FIELD_NUMBER; + hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER; hash = (53 * hash) + getFamilyPathList().hashCode(); } if (hasAssignSeqNum()) { - hash = (37 * hash) + ASSIGNSEQNUM_FIELD_NUMBER; + hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getAssignSeqNum()); } if (hasFsToken()) { - hash = (37 * hash) + FSTOKEN_FIELD_NUMBER; + hash = (37 * hash) + FS_TOKEN_FIELD_NUMBER; hash = (53 * hash) + getFsToken().hashCode(); } if (hasBulkToken()) { - hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER; + hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER; hash = (53 * hash) + getBulkToken().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -597,7 +597,7 @@ public final class SecureBulkLoadProtos { private int bitField0_; - // repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1; + // repeated .BulkLoadHFileRequest.FamilyPath family_path = 1; private java.util.List familyPath_ = java.util.Collections.emptyList(); private void ensureFamilyPathIsMutable() { @@ -783,7 +783,7 @@ public final class SecureBulkLoadProtos { return familyPathBuilder_; } - // optional bool assignSeqNum = 2; + // optional bool assign_seq_num = 2; private boolean assignSeqNum_ ; public boolean hasAssignSeqNum() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -804,7 +804,7 @@ public final class SecureBulkLoadProtos { return this; } - // required .DelegationTokenProto fsToken = 3; + // required .DelegationTokenProto fs_token = 3; private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_; @@ -894,7 +894,7 @@ public final class SecureBulkLoadProtos { return fsTokenBuilder_; } - // required string bulkToken = 4; + // required string bulk_token = 4; private java.lang.Object bulkToken_ = ""; public boolean hasBulkToken() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1972,7 +1972,7 @@ public final class SecureBulkLoadProtos { public interface PrepareBulkLoadRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); } @@ -2005,8 +2005,8 @@ public final class SecureBulkLoadProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2088,7 +2088,7 @@ public final class SecureBulkLoadProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2316,7 +2316,7 @@ public final class SecureBulkLoadProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2354,7 +2354,7 @@ public final class SecureBulkLoadProtos { public interface PrepareBulkLoadResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string bulkToken = 1; + // required string bulk_token = 1; boolean hasBulkToken(); String getBulkToken(); } @@ -2387,8 +2387,8 @@ public final class SecureBulkLoadProtos { } private int bitField0_; - // required string bulkToken = 1; - public static final int BULKTOKEN_FIELD_NUMBER = 1; + // required string bulk_token = 1; + public static final int BULK_TOKEN_FIELD_NUMBER = 1; private java.lang.Object bulkToken_; public boolean hasBulkToken() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2492,7 +2492,7 @@ public final class SecureBulkLoadProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBulkToken()) { - hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER; + hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER; hash = (53 * hash) + getBulkToken().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2720,7 +2720,7 @@ public final class SecureBulkLoadProtos { private int bitField0_; - // required string bulkToken = 1; + // required string bulk_token = 1; private java.lang.Object bulkToken_ = ""; public boolean hasBulkToken() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2770,7 +2770,7 @@ public final class SecureBulkLoadProtos { public interface CleanupBulkLoadRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string bulkToken = 1; + // required string bulk_token = 1; boolean hasBulkToken(); String getBulkToken(); } @@ -2803,8 +2803,8 @@ public final class SecureBulkLoadProtos { } private int bitField0_; - // required string bulkToken = 1; - public static final int BULKTOKEN_FIELD_NUMBER = 1; + // required string bulk_token = 1; + public static final int BULK_TOKEN_FIELD_NUMBER = 1; private java.lang.Object bulkToken_; public boolean hasBulkToken() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2908,7 +2908,7 @@ public final class SecureBulkLoadProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBulkToken()) { - hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER; + hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER; hash = (53 * hash) + getBulkToken().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -3136,7 +3136,7 @@ public final class SecureBulkLoadProtos { private int bitField0_; - // required string bulkToken = 1; + // required string bulk_token = 1; private java.lang.Object bulkToken_ = ""; public boolean hasBulkToken() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -3879,27 +3879,27 @@ public final class SecureBulkLoadProtos { static { java.lang.String[] descriptorData = { "\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" + - "ent.proto\"\244\001\n\033SecureBulkLoadHFilesReques" + - "t\0224\n\nfamilyPath\030\001 \003(\0132 .BulkLoadHFileReq" + - "uest.FamilyPath\022\024\n\014assignSeqNum\030\002 \001(\010\022&\n" + - "\007fsToken\030\003 \002(\0132\025.DelegationTokenProto\022\021\n" + - "\tbulkToken\030\004 \002(\t\".\n\034SecureBulkLoadHFiles" + - "Response\022\016\n\006loaded\030\001 \002(\010\"[\n\024DelegationTo" + - "kenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" + - "\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"+\n" + - "\026PrepareBulkLoadRequest\022\021\n\ttableName\030\001 \002", - "(\014\",\n\027PrepareBulkLoadResponse\022\021\n\tbulkTok" + - "en\030\001 \002(\t\"+\n\026CleanupBulkLoadRequest\022\021\n\tbu" + - "lkToken\030\001 \002(\t\"\031\n\027CleanupBulkLoadResponse" + - "2\370\001\n\025SecureBulkLoadService\022D\n\017prepareBul" + - "kLoad\022\027.PrepareBulkLoadRequest\032\030.Prepare" + - "BulkLoadResponse\022S\n\024secureBulkLoadHFiles" + - "\022\034.SecureBulkLoadHFilesRequest\032\035.SecureB" + - "ulkLoadHFilesResponse\022D\n\017cleanupBulkLoad" + - "\022\027.CleanupBulkLoadRequest\032\030.CleanupBulkL" + - "oadResponseBJ\n*org.apache.hadoop.hbase.p", - "rotobuf.generatedB\024SecureBulkLoadProtosH" + - "\001\210\001\001\240\001\001" + "ent.proto\"\251\001\n\033SecureBulkLoadHFilesReques" + + "t\0225\n\013family_path\030\001 \003(\0132 .BulkLoadHFileRe" + + "quest.FamilyPath\022\026\n\016assign_seq_num\030\002 \001(\010" + + "\022\'\n\010fs_token\030\003 \002(\0132\025.DelegationTokenProt" + + "o\022\022\n\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadH" + + "FilesResponse\022\016\n\006loaded\030\001 \002(\010\"[\n\024Delegat" + + "ionTokenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" + + "sword\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001" + + "(\t\",\n\026PrepareBulkLoadRequest\022\022\n\ntable_na", + "me\030\001 \002(\014\"-\n\027PrepareBulkLoadResponse\022\022\n\nb" + + "ulk_token\030\001 \002(\t\",\n\026CleanupBulkLoadReques" + + "t\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoad" + + "Response2\370\001\n\025SecureBulkLoadService\022D\n\017Pr" + + "epareBulkLoad\022\027.PrepareBulkLoadRequest\032\030" + + ".PrepareBulkLoadResponse\022S\n\024SecureBulkLo" + + "adHFiles\022\034.SecureBulkLoadHFilesRequest\032\035" + + ".SecureBulkLoadHFilesResponse\022D\n\017Cleanup" + + "BulkLoad\022\027.CleanupBulkLoadRequest\032\030.Clea" + + "nupBulkLoadResponseBJ\n*org.apache.hadoop", + ".hbase.protobuf.generatedB\024SecureBulkLoa" + + "dProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/Tracing.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/Tracing.java index e634f17..f5a78ad 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/Tracing.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/Tracing.java @@ -11,11 +11,11 @@ public final class Tracing { public interface RPCTInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional int64 traceId = 1; + // optional int64 trace_id = 1; boolean hasTraceId(); long getTraceId(); - // optional int64 parentId = 2; + // optional int64 parent_id = 2; boolean hasParentId(); long getParentId(); } @@ -48,8 +48,8 @@ public final class Tracing { } private int bitField0_; - // optional int64 traceId = 1; - public static final int TRACEID_FIELD_NUMBER = 1; + // optional int64 trace_id = 1; + public static final int TRACE_ID_FIELD_NUMBER = 1; private long traceId_; public boolean hasTraceId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -58,8 +58,8 @@ public final class Tracing { return traceId_; } - // optional int64 parentId = 2; - public static final int PARENTID_FIELD_NUMBER = 2; + // optional int64 parent_id = 2; + public static final int PARENT_ID_FIELD_NUMBER = 2; private long parentId_; public boolean hasParentId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -150,11 +150,11 @@ public final class Tracing { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTraceId()) { - hash = (37 * hash) + TRACEID_FIELD_NUMBER; + hash = (37 * hash) + TRACE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getTraceId()); } if (hasParentId()) { - hash = (37 * hash) + PARENTID_FIELD_NUMBER; + hash = (37 * hash) + PARENT_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getParentId()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -392,7 +392,7 @@ public final class Tracing { private int bitField0_; - // optional int64 traceId = 1; + // optional int64 trace_id = 1; private long traceId_ ; public boolean hasTraceId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -413,7 +413,7 @@ public final class Tracing { return this; } - // optional int64 parentId = 2; + // optional int64 parent_id = 2; private long parentId_ ; public boolean hasParentId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -459,10 +459,10 @@ public final class Tracing { descriptor; static { java.lang.String[] descriptorData = { - "\n\rTracing.proto\"-\n\010RPCTInfo\022\017\n\007traceId\030\001" + - " \001(\003\022\020\n\010parentId\030\002 \001(\003B:\n*org.apache.had" + - "oop.hbase.protobuf.generatedB\007TracingH\001\240" + - "\001\001" + "\n\rTracing.proto\"/\n\010RPCTInfo\022\020\n\010trace_id\030" + + "\001 \001(\003\022\021\n\tparent_id\030\002 \001(\003B:\n*org.apache.h" + + "adoop.hbase.protobuf.generatedB\007TracingH" + + "\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java index f72c90e..338f324 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java @@ -80,7 +80,7 @@ public final class WALProtos { public interface WALHeaderOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bool hasCompression = 1; + // optional bool has_compression = 1; boolean hasHasCompression(); boolean getHasCompression(); } @@ -113,8 +113,8 @@ public final class WALProtos { } private int bitField0_; - // optional bool hasCompression = 1; - public static final int HASCOMPRESSION_FIELD_NUMBER = 1; + // optional bool has_compression = 1; + public static final int HAS_COMPRESSION_FIELD_NUMBER = 1; private boolean hasCompression_; public boolean hasHasCompression() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -192,7 +192,7 @@ public final class WALProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasHasCompression()) { - hash = (37 * hash) + HASCOMPRESSION_FIELD_NUMBER; + hash = (37 * hash) + HAS_COMPRESSION_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getHasCompression()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -416,7 +416,7 @@ public final class WALProtos { private int bitField0_; - // optional bool hasCompression = 1; + // optional bool has_compression = 1; private boolean hasCompression_ ; public boolean hasHasCompression() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -451,23 +451,23 @@ public final class WALProtos { public interface WALKeyOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes encodedRegionName = 1; + // required bytes encoded_region_name = 1; boolean hasEncodedRegionName(); com.google.protobuf.ByteString getEncodedRegionName(); - // required bytes tableName = 2; + // required bytes table_name = 2; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // required uint64 logSequenceNumber = 3; + // required uint64 log_sequence_number = 3; boolean hasLogSequenceNumber(); long getLogSequenceNumber(); - // required uint64 writeTime = 4; + // required uint64 write_time = 4; boolean hasWriteTime(); long getWriteTime(); - // optional .UUID clusterId = 5; + // optional .UUID cluster_id = 5; boolean hasClusterId(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder(); @@ -482,7 +482,7 @@ public final class WALProtos { org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder( int index); - // optional uint32 followingKvCount = 7; + // optional uint32 following_kv_count = 7; boolean hasFollowingKvCount(); int getFollowingKvCount(); } @@ -515,8 +515,8 @@ public final class WALProtos { } private int bitField0_; - // required bytes encodedRegionName = 1; - public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 1; + // required bytes encoded_region_name = 1; + public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString encodedRegionName_; public boolean hasEncodedRegionName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -525,8 +525,8 @@ public final class WALProtos { return encodedRegionName_; } - // required bytes tableName = 2; - public static final int TABLENAME_FIELD_NUMBER = 2; + // required bytes table_name = 2; + public static final int TABLE_NAME_FIELD_NUMBER = 2; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -535,8 +535,8 @@ public final class WALProtos { return tableName_; } - // required uint64 logSequenceNumber = 3; - public static final int LOGSEQUENCENUMBER_FIELD_NUMBER = 3; + // required uint64 log_sequence_number = 3; + public static final int LOG_SEQUENCE_NUMBER_FIELD_NUMBER = 3; private long logSequenceNumber_; public boolean hasLogSequenceNumber() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -545,8 +545,8 @@ public final class WALProtos { return logSequenceNumber_; } - // required uint64 writeTime = 4; - public static final int WRITETIME_FIELD_NUMBER = 4; + // required uint64 write_time = 4; + public static final int WRITE_TIME_FIELD_NUMBER = 4; private long writeTime_; public boolean hasWriteTime() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -555,8 +555,8 @@ public final class WALProtos { return writeTime_; } - // optional .UUID clusterId = 5; - public static final int CLUSTERID_FIELD_NUMBER = 5; + // optional .UUID cluster_id = 5; + public static final int CLUSTER_ID_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_; public boolean hasClusterId() { return ((bitField0_ & 0x00000010) == 0x00000010); @@ -589,8 +589,8 @@ public final class WALProtos { return scopes_.get(index); } - // optional uint32 followingKvCount = 7; - public static final int FOLLOWINGKVCOUNT_FIELD_NUMBER = 7; + // optional uint32 following_kv_count = 7; + public static final int FOLLOWING_KV_COUNT_FIELD_NUMBER = 7; private int followingKvCount_; public boolean hasFollowingKvCount() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -771,23 +771,23 @@ public final class WALProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasEncodedRegionName()) { - hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER; + hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER; hash = (53 * hash) + getEncodedRegionName().hashCode(); } if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasLogSequenceNumber()) { - hash = (37 * hash) + LOGSEQUENCENUMBER_FIELD_NUMBER; + hash = (37 * hash) + LOG_SEQUENCE_NUMBER_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLogSequenceNumber()); } if (hasWriteTime()) { - hash = (37 * hash) + WRITETIME_FIELD_NUMBER; + hash = (37 * hash) + WRITE_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getWriteTime()); } if (hasClusterId()) { - hash = (37 * hash) + CLUSTERID_FIELD_NUMBER; + hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER; hash = (53 * hash) + getClusterId().hashCode(); } if (getScopesCount() > 0) { @@ -795,7 +795,7 @@ public final class WALProtos { hash = (53 * hash) + getScopesList().hashCode(); } if (hasFollowingKvCount()) { - hash = (37 * hash) + FOLLOWINGKVCOUNT_FIELD_NUMBER; + hash = (37 * hash) + FOLLOWING_KV_COUNT_FIELD_NUMBER; hash = (53 * hash) + getFollowingKvCount(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1178,7 +1178,7 @@ public final class WALProtos { private int bitField0_; - // required bytes encodedRegionName = 1; + // required bytes encoded_region_name = 1; private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasEncodedRegionName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1202,7 +1202,7 @@ public final class WALProtos { return this; } - // required bytes tableName = 2; + // required bytes table_name = 2; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1226,7 +1226,7 @@ public final class WALProtos { return this; } - // required uint64 logSequenceNumber = 3; + // required uint64 log_sequence_number = 3; private long logSequenceNumber_ ; public boolean hasLogSequenceNumber() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1247,7 +1247,7 @@ public final class WALProtos { return this; } - // required uint64 writeTime = 4; + // required uint64 write_time = 4; private long writeTime_ ; public boolean hasWriteTime() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1268,7 +1268,7 @@ public final class WALProtos { return this; } - // optional .UUID clusterId = 5; + // optional .UUID cluster_id = 5; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_; @@ -1544,7 +1544,7 @@ public final class WALProtos { return scopesBuilder_; } - // optional uint32 followingKvCount = 7; + // optional uint32 following_kv_count = 7; private int followingKvCount_ ; public boolean hasFollowingKvCount() { return ((bitField0_ & 0x00000040) == 0x00000040); @@ -1583,7 +1583,7 @@ public final class WALProtos { boolean hasFamily(); com.google.protobuf.ByteString getFamily(); - // required .ScopeType scopeType = 2; + // required .ScopeType scope_type = 2; boolean hasScopeType(); org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType(); } @@ -1626,8 +1626,8 @@ public final class WALProtos { return family_; } - // required .ScopeType scopeType = 2; - public static final int SCOPETYPE_FIELD_NUMBER = 2; + // required .ScopeType scope_type = 2; + public static final int SCOPE_TYPE_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_; public boolean hasScopeType() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1730,7 +1730,7 @@ public final class WALProtos { hash = (53 * hash) + getFamily().hashCode(); } if (hasScopeType()) { - hash = (37 * hash) + SCOPETYPE_FIELD_NUMBER; + hash = (37 * hash) + SCOPE_TYPE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getScopeType()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2006,7 +2006,7 @@ public final class WALProtos { return this; } - // required .ScopeType scopeType = 2; + // required .ScopeType scope_type = 2; private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL; public boolean hasScopeType() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2044,29 +2044,29 @@ public final class WALProtos { public interface CompactionDescriptorOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes tableName = 1; + // required bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // required bytes encodedRegionName = 2; + // required bytes encoded_region_name = 2; boolean hasEncodedRegionName(); com.google.protobuf.ByteString getEncodedRegionName(); - // required bytes familyName = 3; + // required bytes family_name = 3; boolean hasFamilyName(); com.google.protobuf.ByteString getFamilyName(); - // repeated string compactionInput = 4; + // repeated string compaction_input = 4; java.util.List getCompactionInputList(); int getCompactionInputCount(); String getCompactionInput(int index); - // repeated string compactionOutput = 5; + // repeated string compaction_output = 5; java.util.List getCompactionOutputList(); int getCompactionOutputCount(); String getCompactionOutput(int index); - // required string storeHomeDir = 6; + // required string store_home_dir = 6; boolean hasStoreHomeDir(); String getStoreHomeDir(); } @@ -2099,8 +2099,8 @@ public final class WALProtos { } private int bitField0_; - // required bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // required bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2109,8 +2109,8 @@ public final class WALProtos { return tableName_; } - // required bytes encodedRegionName = 2; - public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 2; + // required bytes encoded_region_name = 2; + public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 2; private com.google.protobuf.ByteString encodedRegionName_; public boolean hasEncodedRegionName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2119,8 +2119,8 @@ public final class WALProtos { return encodedRegionName_; } - // required bytes familyName = 3; - public static final int FAMILYNAME_FIELD_NUMBER = 3; + // required bytes family_name = 3; + public static final int FAMILY_NAME_FIELD_NUMBER = 3; private com.google.protobuf.ByteString familyName_; public boolean hasFamilyName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2129,8 +2129,8 @@ public final class WALProtos { return familyName_; } - // repeated string compactionInput = 4; - public static final int COMPACTIONINPUT_FIELD_NUMBER = 4; + // repeated string compaction_input = 4; + public static final int COMPACTION_INPUT_FIELD_NUMBER = 4; private com.google.protobuf.LazyStringList compactionInput_; public java.util.List getCompactionInputList() { @@ -2143,8 +2143,8 @@ public final class WALProtos { return compactionInput_.get(index); } - // repeated string compactionOutput = 5; - public static final int COMPACTIONOUTPUT_FIELD_NUMBER = 5; + // repeated string compaction_output = 5; + public static final int COMPACTION_OUTPUT_FIELD_NUMBER = 5; private com.google.protobuf.LazyStringList compactionOutput_; public java.util.List getCompactionOutputList() { @@ -2157,8 +2157,8 @@ public final class WALProtos { return compactionOutput_.get(index); } - // required string storeHomeDir = 6; - public static final int STOREHOMEDIR_FIELD_NUMBER = 6; + // required string store_home_dir = 6; + public static final int STORE_HOME_DIR_FIELD_NUMBER = 6; private java.lang.Object storeHomeDir_; public boolean hasStoreHomeDir() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -2343,27 +2343,27 @@ public final class WALProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasEncodedRegionName()) { - hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER; + hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER; hash = (53 * hash) + getEncodedRegionName().hashCode(); } if (hasFamilyName()) { - hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER; + hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER; hash = (53 * hash) + getFamilyName().hashCode(); } if (getCompactionInputCount() > 0) { - hash = (37 * hash) + COMPACTIONINPUT_FIELD_NUMBER; + hash = (37 * hash) + COMPACTION_INPUT_FIELD_NUMBER; hash = (53 * hash) + getCompactionInputList().hashCode(); } if (getCompactionOutputCount() > 0) { - hash = (37 * hash) + COMPACTIONOUTPUT_FIELD_NUMBER; + hash = (37 * hash) + COMPACTION_OUTPUT_FIELD_NUMBER; hash = (53 * hash) + getCompactionOutputList().hashCode(); } if (hasStoreHomeDir()) { - hash = (37 * hash) + STOREHOMEDIR_FIELD_NUMBER; + hash = (37 * hash) + STORE_HOME_DIR_FIELD_NUMBER; hash = (53 * hash) + getStoreHomeDir().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2691,7 +2691,7 @@ public final class WALProtos { private int bitField0_; - // required bytes tableName = 1; + // required bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2715,7 +2715,7 @@ public final class WALProtos { return this; } - // required bytes encodedRegionName = 2; + // required bytes encoded_region_name = 2; private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasEncodedRegionName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2739,7 +2739,7 @@ public final class WALProtos { return this; } - // required bytes familyName = 3; + // required bytes family_name = 3; private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasFamilyName() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2763,7 +2763,7 @@ public final class WALProtos { return this; } - // repeated string compactionInput = 4; + // repeated string compaction_input = 4; private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureCompactionInputIsMutable() { if (!((bitField0_ & 0x00000008) == 0x00000008)) { @@ -2819,7 +2819,7 @@ public final class WALProtos { onChanged(); } - // repeated string compactionOutput = 5; + // repeated string compaction_output = 5; private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureCompactionOutputIsMutable() { if (!((bitField0_ & 0x00000010) == 0x00000010)) { @@ -2875,7 +2875,7 @@ public final class WALProtos { onChanged(); } - // required string storeHomeDir = 6; + // required string store_home_dir = 6; private java.lang.Object storeHomeDir_ = ""; public boolean hasStoreHomeDir() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -3256,22 +3256,23 @@ public final class WALProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\tWAL.proto\032\013hbase.proto\"#\n\tWALHeader\022\026\n" + - "\016hasCompression\030\001 \001(\010\"\266\001\n\006WALKey\022\031\n\021enco" + - "dedRegionName\030\001 \002(\014\022\021\n\ttableName\030\002 \002(\014\022\031" + - "\n\021logSequenceNumber\030\003 \002(\004\022\021\n\twriteTime\030\004" + - " \002(\004\022\030\n\tclusterId\030\005 \001(\0132\005.UUID\022\034\n\006scopes" + - "\030\006 \003(\0132\014.FamilyScope\022\030\n\020followingKvCount" + - "\030\007 \001(\r\"<\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\035\n" + - "\tscopeType\030\002 \002(\0162\n.ScopeType\"\241\001\n\024Compact" + - "ionDescriptor\022\021\n\ttableName\030\001 \002(\014\022\031\n\021enco" + - "dedRegionName\030\002 \002(\014\022\022\n\nfamilyName\030\003 \002(\014\022", - "\027\n\017compactionInput\030\004 \003(\t\022\030\n\020compactionOu" + - "tput\030\005 \003(\t\022\024\n\014storeHomeDir\030\006 \002(\t\"\014\n\nWALT" + - "railer*F\n\tScopeType\022\033\n\027REPLICATION_SCOPE" + - "_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001B?" + - "\n*org.apache.hadoop.hbase.protobuf.gener" + - "atedB\tWALProtosH\001\210\001\000\240\001\001" + "\n\tWAL.proto\032\013hbase.proto\"$\n\tWALHeader\022\027\n" + + "\017has_compression\030\001 \001(\010\"\277\001\n\006WALKey\022\033\n\023enc" + + "oded_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002" + + "(\014\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite" + + "_time\030\004 \002(\004\022\031\n\ncluster_id\030\005 \001(\0132\005.UUID\022\034" + + "\n\006scopes\030\006 \003(\0132\014.FamilyScope\022\032\n\022followin" + + "g_kv_count\030\007 \001(\r\"=\n\013FamilyScope\022\016\n\006famil" + + "y\030\001 \002(\014\022\036\n\nscope_type\030\002 \002(\0162\n.ScopeType\"" + + "\251\001\n\024CompactionDescriptor\022\022\n\ntable_name\030\001" + + " \002(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013fam", + "ily_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t" + + "\022\031\n\021compaction_output\030\005 \003(\t\022\026\n\016store_hom" + + "e_dir\030\006 \002(\t\"\014\n\nWALTrailer*F\n\tScopeType\022\033" + + "\n\027REPLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATI" + + "ON_SCOPE_GLOBAL\020\001B?\n*org.apache.hadoop.h" + + "base.protobuf.generatedB\tWALProtosH\001\210\001\000\240" + + "\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java index 4d7160a..db01207 100644 --- hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java +++ hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java @@ -16,7 +16,7 @@ public final class ZooKeeperProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); - // optional uint32 rpcVersion = 2; + // optional uint32 rpc_version = 2; boolean hasRpcVersion(); int getRpcVersion(); } @@ -62,8 +62,8 @@ public final class ZooKeeperProtos { return server_; } - // optional uint32 rpcVersion = 2; - public static final int RPCVERSION_FIELD_NUMBER = 2; + // optional uint32 rpc_version = 2; + public static final int RPC_VERSION_FIELD_NUMBER = 2; private int rpcVersion_; public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -166,7 +166,7 @@ public final class ZooKeeperProtos { hash = (53 * hash) + getServer().hashCode(); } if (hasRpcVersion()) { - hash = (37 * hash) + RPCVERSION_FIELD_NUMBER; + hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER; hash = (53 * hash) + getRpcVersion(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -515,7 +515,7 @@ public final class ZooKeeperProtos { return serverBuilder_; } - // optional uint32 rpcVersion = 2; + // optional uint32 rpc_version = 2; private int rpcVersion_ ; public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -555,7 +555,7 @@ public final class ZooKeeperProtos { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder(); - // optional uint32 rpcVersion = 2; + // optional uint32 rpc_version = 2; boolean hasRpcVersion(); int getRpcVersion(); } @@ -601,8 +601,8 @@ public final class ZooKeeperProtos { return master_; } - // optional uint32 rpcVersion = 2; - public static final int RPCVERSION_FIELD_NUMBER = 2; + // optional uint32 rpc_version = 2; + public static final int RPC_VERSION_FIELD_NUMBER = 2; private int rpcVersion_; public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -705,7 +705,7 @@ public final class ZooKeeperProtos { hash = (53 * hash) + getMaster().hashCode(); } if (hasRpcVersion()) { - hash = (37 * hash) + RPCVERSION_FIELD_NUMBER; + hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER; hash = (53 * hash) + getRpcVersion(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1054,7 +1054,7 @@ public final class ZooKeeperProtos { return masterBuilder_; } - // optional uint32 rpcVersion = 2; + // optional uint32 rpc_version = 2; private int rpcVersion_ ; public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1089,7 +1089,7 @@ public final class ZooKeeperProtos { public interface ClusterUpOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string startDate = 1; + // required string start_date = 1; boolean hasStartDate(); String getStartDate(); } @@ -1122,8 +1122,8 @@ public final class ZooKeeperProtos { } private int bitField0_; - // required string startDate = 1; - public static final int STARTDATE_FIELD_NUMBER = 1; + // required string start_date = 1; + public static final int START_DATE_FIELD_NUMBER = 1; private java.lang.Object startDate_; public boolean hasStartDate() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1227,7 +1227,7 @@ public final class ZooKeeperProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasStartDate()) { - hash = (37 * hash) + STARTDATE_FIELD_NUMBER; + hash = (37 * hash) + START_DATE_FIELD_NUMBER; hash = (53 * hash) + getStartDate().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -1455,7 +1455,7 @@ public final class ZooKeeperProtos { private int bitField0_; - // required string startDate = 1; + // required string start_date = 1; private java.lang.Object startDate_ = ""; public boolean hasStartDate() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1505,19 +1505,19 @@ public final class ZooKeeperProtos { public interface RegionTransitionOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required uint32 eventTypeCode = 1; + // required uint32 event_type_code = 1; boolean hasEventTypeCode(); int getEventTypeCode(); - // required bytes regionName = 2; + // required bytes region_name = 2; boolean hasRegionName(); com.google.protobuf.ByteString getRegionName(); - // required uint64 createTime = 3; + // required uint64 create_time = 3; boolean hasCreateTime(); long getCreateTime(); - // required .ServerName serverName = 4; + // required .ServerName server_name = 4; boolean hasServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); @@ -1555,8 +1555,8 @@ public final class ZooKeeperProtos { } private int bitField0_; - // required uint32 eventTypeCode = 1; - public static final int EVENTTYPECODE_FIELD_NUMBER = 1; + // required uint32 event_type_code = 1; + public static final int EVENT_TYPE_CODE_FIELD_NUMBER = 1; private int eventTypeCode_; public boolean hasEventTypeCode() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -1565,8 +1565,8 @@ public final class ZooKeeperProtos { return eventTypeCode_; } - // required bytes regionName = 2; - public static final int REGIONNAME_FIELD_NUMBER = 2; + // required bytes region_name = 2; + public static final int REGION_NAME_FIELD_NUMBER = 2; private com.google.protobuf.ByteString regionName_; public boolean hasRegionName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -1575,8 +1575,8 @@ public final class ZooKeeperProtos { return regionName_; } - // required uint64 createTime = 3; - public static final int CREATETIME_FIELD_NUMBER = 3; + // required uint64 create_time = 3; + public static final int CREATE_TIME_FIELD_NUMBER = 3; private long createTime_; public boolean hasCreateTime() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -1585,8 +1585,8 @@ public final class ZooKeeperProtos { return createTime_; } - // required .ServerName serverName = 4; - public static final int SERVERNAME_FIELD_NUMBER = 4; + // required .ServerName server_name = 4; + public static final int SERVER_NAME_FIELD_NUMBER = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; public boolean hasServerName() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -1749,19 +1749,19 @@ public final class ZooKeeperProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasEventTypeCode()) { - hash = (37 * hash) + EVENTTYPECODE_FIELD_NUMBER; + hash = (37 * hash) + EVENT_TYPE_CODE_FIELD_NUMBER; hash = (53 * hash) + getEventTypeCode(); } if (hasRegionName()) { - hash = (37 * hash) + REGIONNAME_FIELD_NUMBER; + hash = (37 * hash) + REGION_NAME_FIELD_NUMBER; hash = (53 * hash) + getRegionName().hashCode(); } if (hasCreateTime()) { - hash = (37 * hash) + CREATETIME_FIELD_NUMBER; + hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCreateTime()); } if (hasServerName()) { - hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; + hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } if (hasPayload()) { @@ -2078,7 +2078,7 @@ public final class ZooKeeperProtos { private int bitField0_; - // required uint32 eventTypeCode = 1; + // required uint32 event_type_code = 1; private int eventTypeCode_ ; public boolean hasEventTypeCode() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -2099,7 +2099,7 @@ public final class ZooKeeperProtos { return this; } - // required bytes regionName = 2; + // required bytes region_name = 2; private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasRegionName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2123,7 +2123,7 @@ public final class ZooKeeperProtos { return this; } - // required uint64 createTime = 3; + // required uint64 create_time = 3; private long createTime_ ; public boolean hasCreateTime() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -2144,7 +2144,7 @@ public final class ZooKeeperProtos { return this; } - // required .ServerName serverName = 4; + // required .ServerName server_name = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; @@ -2276,7 +2276,7 @@ public final class ZooKeeperProtos { boolean hasState(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState(); - // required .ServerName serverName = 2; + // required .ServerName server_name = 2; boolean hasServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); @@ -2398,8 +2398,8 @@ public final class ZooKeeperProtos { return state_; } - // required .ServerName serverName = 2; - public static final int SERVERNAME_FIELD_NUMBER = 2; + // required .ServerName server_name = 2; + public static final int SERVER_NAME_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; public boolean hasServerName() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -2509,7 +2509,7 @@ public final class ZooKeeperProtos { hash = (53 * hash) + hashEnum(getState()); } if (hasServerName()) { - hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; + hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -2802,7 +2802,7 @@ public final class ZooKeeperProtos { return this; } - // required .ServerName serverName = 2; + // required .ServerName server_name = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; @@ -4621,7 +4621,7 @@ public final class ZooKeeperProtos { public interface ReplicationLockOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required string lockOwner = 1; + // required string lock_owner = 1; boolean hasLockOwner(); String getLockOwner(); } @@ -4654,8 +4654,8 @@ public final class ZooKeeperProtos { } private int bitField0_; - // required string lockOwner = 1; - public static final int LOCKOWNER_FIELD_NUMBER = 1; + // required string lock_owner = 1; + public static final int LOCK_OWNER_FIELD_NUMBER = 1; private java.lang.Object lockOwner_; public boolean hasLockOwner() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -4759,7 +4759,7 @@ public final class ZooKeeperProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLockOwner()) { - hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER; + hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER; hash = (53 * hash) + getLockOwner().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -4987,7 +4987,7 @@ public final class ZooKeeperProtos { private int bitField0_; - // required string lockOwner = 1; + // required string lock_owner = 1; private java.lang.Object lockOwner_ = ""; public boolean hasLockOwner() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -5037,20 +5037,20 @@ public final class ZooKeeperProtos { public interface TableLockOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional bytes tableName = 1; + // optional bytes table_name = 1; boolean hasTableName(); com.google.protobuf.ByteString getTableName(); - // optional .ServerName lockOwner = 2; + // optional .ServerName lock_owner = 2; boolean hasLockOwner(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder(); - // optional int64 threadId = 3; + // optional int64 thread_id = 3; boolean hasThreadId(); long getThreadId(); - // optional bool isShared = 4; + // optional bool is_shared = 4; boolean hasIsShared(); boolean getIsShared(); @@ -5058,7 +5058,7 @@ public final class ZooKeeperProtos { boolean hasPurpose(); String getPurpose(); - // optional int64 createTime = 6; + // optional int64 create_time = 6; boolean hasCreateTime(); long getCreateTime(); } @@ -5091,8 +5091,8 @@ public final class ZooKeeperProtos { } private int bitField0_; - // optional bytes tableName = 1; - public static final int TABLENAME_FIELD_NUMBER = 1; + // optional bytes table_name = 1; + public static final int TABLE_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString tableName_; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -5101,8 +5101,8 @@ public final class ZooKeeperProtos { return tableName_; } - // optional .ServerName lockOwner = 2; - public static final int LOCKOWNER_FIELD_NUMBER = 2; + // optional .ServerName lock_owner = 2; + public static final int LOCK_OWNER_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_; public boolean hasLockOwner() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -5114,8 +5114,8 @@ public final class ZooKeeperProtos { return lockOwner_; } - // optional int64 threadId = 3; - public static final int THREADID_FIELD_NUMBER = 3; + // optional int64 thread_id = 3; + public static final int THREAD_ID_FIELD_NUMBER = 3; private long threadId_; public boolean hasThreadId() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -5124,8 +5124,8 @@ public final class ZooKeeperProtos { return threadId_; } - // optional bool isShared = 4; - public static final int ISSHARED_FIELD_NUMBER = 4; + // optional bool is_shared = 4; + public static final int IS_SHARED_FIELD_NUMBER = 4; private boolean isShared_; public boolean hasIsShared() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -5166,8 +5166,8 @@ public final class ZooKeeperProtos { } } - // optional int64 createTime = 6; - public static final int CREATETIME_FIELD_NUMBER = 6; + // optional int64 create_time = 6; + public static final int CREATE_TIME_FIELD_NUMBER = 6; private long createTime_; public boolean hasCreateTime() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -5316,19 +5316,19 @@ public final class ZooKeeperProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { - hash = (37 * hash) + TABLENAME_FIELD_NUMBER; + hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasLockOwner()) { - hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER; + hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER; hash = (53 * hash) + getLockOwner().hashCode(); } if (hasThreadId()) { - hash = (37 * hash) + THREADID_FIELD_NUMBER; + hash = (37 * hash) + THREAD_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getThreadId()); } if (hasIsShared()) { - hash = (37 * hash) + ISSHARED_FIELD_NUMBER; + hash = (37 * hash) + IS_SHARED_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getIsShared()); } if (hasPurpose()) { @@ -5336,7 +5336,7 @@ public final class ZooKeeperProtos { hash = (53 * hash) + getPurpose().hashCode(); } if (hasCreateTime()) { - hash = (37 * hash) + CREATETIME_FIELD_NUMBER; + hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCreateTime()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -5649,7 +5649,7 @@ public final class ZooKeeperProtos { private int bitField0_; - // optional bytes tableName = 1; + // optional bytes table_name = 1; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -5673,7 +5673,7 @@ public final class ZooKeeperProtos { return this; } - // optional .ServerName lockOwner = 2; + // optional .ServerName lock_owner = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_; @@ -5763,7 +5763,7 @@ public final class ZooKeeperProtos { return lockOwnerBuilder_; } - // optional int64 threadId = 3; + // optional int64 thread_id = 3; private long threadId_ ; public boolean hasThreadId() { return ((bitField0_ & 0x00000004) == 0x00000004); @@ -5784,7 +5784,7 @@ public final class ZooKeeperProtos { return this; } - // optional bool isShared = 4; + // optional bool is_shared = 4; private boolean isShared_ ; public boolean hasIsShared() { return ((bitField0_ & 0x00000008) == 0x00000008); @@ -5841,7 +5841,7 @@ public final class ZooKeeperProtos { onChanged(); } - // optional int64 createTime = 6; + // optional int64 create_time = 6; private long createTime_ ; public boolean hasCreateTime() { return ((bitField0_ & 0x00000020) == 0x00000020); @@ -5876,11 +5876,11 @@ public final class ZooKeeperProtos { public interface StoreSequenceIdOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bytes familyName = 1; + // required bytes family_name = 1; boolean hasFamilyName(); com.google.protobuf.ByteString getFamilyName(); - // required uint64 sequenceId = 2; + // required uint64 sequence_id = 2; boolean hasSequenceId(); long getSequenceId(); } @@ -5913,8 +5913,8 @@ public final class ZooKeeperProtos { } private int bitField0_; - // required bytes familyName = 1; - public static final int FAMILYNAME_FIELD_NUMBER = 1; + // required bytes family_name = 1; + public static final int FAMILY_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString familyName_; public boolean hasFamilyName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -5923,8 +5923,8 @@ public final class ZooKeeperProtos { return familyName_; } - // required uint64 sequenceId = 2; - public static final int SEQUENCEID_FIELD_NUMBER = 2; + // required uint64 sequence_id = 2; + public static final int SEQUENCE_ID_FIELD_NUMBER = 2; private long sequenceId_; public boolean hasSequenceId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -6023,11 +6023,11 @@ public final class ZooKeeperProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasFamilyName()) { - hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER; + hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER; hash = (53 * hash) + getFamilyName().hashCode(); } if (hasSequenceId()) { - hash = (37 * hash) + SEQUENCEID_FIELD_NUMBER; + hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSequenceId()); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -6273,7 +6273,7 @@ public final class ZooKeeperProtos { private int bitField0_; - // required bytes familyName = 1; + // required bytes family_name = 1; private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY; public boolean hasFamilyName() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6297,7 +6297,7 @@ public final class ZooKeeperProtos { return this; } - // required uint64 sequenceId = 2; + // required uint64 sequence_id = 2; private long sequenceId_ ; public boolean hasSequenceId() { return ((bitField0_ & 0x00000002) == 0x00000002); @@ -6332,11 +6332,11 @@ public final class ZooKeeperProtos { public interface RegionStoreSequenceIdsOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required uint64 lastFlushedSequenceId = 1; + // required uint64 last_flushed_sequence_id = 1; boolean hasLastFlushedSequenceId(); long getLastFlushedSequenceId(); - // repeated .StoreSequenceId storeSequenceId = 2; + // repeated .StoreSequenceId store_sequence_id = 2; java.util.List getStoreSequenceIdList(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index); @@ -6375,8 +6375,8 @@ public final class ZooKeeperProtos { } private int bitField0_; - // required uint64 lastFlushedSequenceId = 1; - public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1; + // required uint64 last_flushed_sequence_id = 1; + public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1; private long lastFlushedSequenceId_; public boolean hasLastFlushedSequenceId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6385,8 +6385,8 @@ public final class ZooKeeperProtos { return lastFlushedSequenceId_; } - // repeated .StoreSequenceId storeSequenceId = 2; - public static final int STORESEQUENCEID_FIELD_NUMBER = 2; + // repeated .StoreSequenceId store_sequence_id = 2; + public static final int STORE_SEQUENCE_ID_FIELD_NUMBER = 2; private java.util.List storeSequenceId_; public java.util.List getStoreSequenceIdList() { return storeSequenceId_; @@ -6495,11 +6495,11 @@ public final class ZooKeeperProtos { int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLastFlushedSequenceId()) { - hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER; + hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLastFlushedSequenceId()); } if (getStoreSequenceIdCount() > 0) { - hash = (37 * hash) + STORESEQUENCEID_FIELD_NUMBER; + hash = (37 * hash) + STORE_SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + getStoreSequenceIdList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); @@ -6781,7 +6781,7 @@ public final class ZooKeeperProtos { private int bitField0_; - // required uint64 lastFlushedSequenceId = 1; + // required uint64 last_flushed_sequence_id = 1; private long lastFlushedSequenceId_ ; public boolean hasLastFlushedSequenceId() { return ((bitField0_ & 0x00000001) == 0x00000001); @@ -6802,7 +6802,7 @@ public final class ZooKeeperProtos { return this; } - // repeated .StoreSequenceId storeSequenceId = 2; + // repeated .StoreSequenceId store_sequence_id = 2; private java.util.List storeSequenceId_ = java.util.Collections.emptyList(); private void ensureStoreSequenceIdIsMutable() { @@ -7073,36 +7073,36 @@ public final class ZooKeeperProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\017ZooKeeper.proto\032\013hbase.proto\"C\n\020MetaRe" + + "\n\017ZooKeeper.proto\032\013hbase.proto\"D\n\020MetaRe" + "gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" + - "\022\n\nrpcVersion\030\002 \001(\r\"9\n\006Master\022\033\n\006master\030" + - "\001 \002(\0132\013.ServerName\022\022\n\nrpcVersion\030\002 \001(\r\"\036" + - "\n\tClusterUp\022\021\n\tstartDate\030\001 \002(\t\"\203\001\n\020Regio" + - "nTransition\022\025\n\reventTypeCode\030\001 \002(\r\022\022\n\nre" + - "gionName\030\002 \002(\014\022\022\n\ncreateTime\030\003 \002(\004\022\037\n\nse" + - "rverName\030\004 \002(\0132\013.ServerName\022\017\n\007payload\030\005" + - " \001(\014\"\230\001\n\014SplitLogTask\022\"\n\005state\030\001 \002(\0162\023.S" + - "plitLogTask.State\022\037\n\nserverName\030\002 \002(\0132\013.", - "ServerName\"C\n\005State\022\016\n\nUNASSIGNED\020\000\022\t\n\005O" + - "WNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\"" + - "n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table.State:\007E" + - "NABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED" + - "\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"%\n\017Repli" + - "cationPeer\022\022\n\nclusterkey\030\001 \002(\t\"^\n\020Replic" + - "ationState\022&\n\005state\030\001 \002(\0162\027.ReplicationS" + - "tate.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISA" + - "BLED\020\001\"+\n\027ReplicationHLogPosition\022\020\n\010pos" + - "ition\030\001 \002(\003\"$\n\017ReplicationLock\022\021\n\tlockOw", - "ner\030\001 \002(\t\"\207\001\n\tTableLock\022\021\n\ttableName\030\001 \001" + - "(\014\022\036\n\tlockOwner\030\002 \001(\0132\013.ServerName\022\020\n\010th" + - "readId\030\003 \001(\003\022\020\n\010isShared\030\004 \001(\010\022\017\n\007purpos" + - "e\030\005 \001(\t\022\022\n\ncreateTime\030\006 \001(\003\"9\n\017StoreSequ" + - "enceId\022\022\n\nfamilyName\030\001 \002(\014\022\022\n\nsequenceId" + - "\030\002 \002(\004\"b\n\026RegionStoreSequenceIds\022\035\n\025last" + - "FlushedSequenceId\030\001 \002(\004\022)\n\017storeSequence" + - "Id\030\002 \003(\0132\020.StoreSequenceIdBE\n*org.apache" + - ".hadoop.hbase.protobuf.generatedB\017ZooKee" + - "perProtosH\001\210\001\001\240\001\001" + "\023\n\013rpc_version\030\002 \001(\r\":\n\006Master\022\033\n\006master" + + "\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(\r" + + "\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\210\001\n\020Re" + + "gionTransition\022\027\n\017event_type_code\030\001 \002(\r\022" + + "\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(" + + "\004\022 \n\013server_name\030\004 \002(\0132\013.ServerName\022\017\n\007p" + + "ayload\030\005 \001(\014\"\231\001\n\014SplitLogTask\022\"\n\005state\030\001" + + " \002(\0162\023.SplitLogTask.State\022 \n\013server_name", + "\030\002 \002(\0132\013.ServerName\"C\n\005State\022\016\n\nUNASSIGN" + + "ED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022" + + "\007\n\003ERR\020\004\"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table" + + ".State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n" + + "\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003" + + "\"%\n\017ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\"" + + "^\n\020ReplicationState\022&\n\005state\030\001 \002(\0162\027.Rep" + + "licationState.State\"\"\n\005State\022\013\n\007ENABLED\020" + + "\000\022\014\n\010DISABLED\020\001\"+\n\027ReplicationHLogPositi" + + "on\022\020\n\010position\030\001 \002(\003\"%\n\017ReplicationLock\022", + "\022\n\nlock_owner\030\001 \002(\t\"\214\001\n\tTableLock\022\022\n\ntab" + + "le_name\030\001 \001(\014\022\037\n\nlock_owner\030\002 \001(\0132\013.Serv" + + "erName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030\004" + + " \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 \001" + + "(\003\";\n\017StoreSequenceId\022\023\n\013family_name\030\001 \002" + + "(\014\022\023\n\013sequence_id\030\002 \002(\004\"g\n\026RegionStoreSe" + + "quenceIds\022 \n\030last_flushed_sequence_id\030\001 " + + "\002(\004\022+\n\021store_sequence_id\030\002 \003(\0132\020.StoreSe" + + "quenceIdBE\n*org.apache.hadoop.hbase.prot" + + "obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { diff --git hbase-protocol/src/main/protobuf/AccessControl.proto hbase-protocol/src/main/protobuf/AccessControl.proto index 4441092..e866993 100644 --- hbase-protocol/src/main/protobuf/AccessControl.proto +++ hbase-protocol/src/main/protobuf/AccessControl.proto @@ -85,15 +85,15 @@ message CheckPermissionsResponse { } service AccessControlService { - rpc grant(GrantRequest) + rpc Grant(GrantRequest) returns (GrantResponse); - rpc revoke(RevokeRequest) + rpc Revoke(RevokeRequest) returns (RevokeResponse); - rpc getUserPermissions(UserPermissionsRequest) + rpc GetUserPermissions(UserPermissionsRequest) returns (UserPermissionsResponse); - rpc checkPermissions(CheckPermissionsRequest) + rpc CheckPermissions(CheckPermissionsRequest) returns (CheckPermissionsResponse); } diff --git hbase-protocol/src/main/protobuf/Admin.proto hbase-protocol/src/main/protobuf/Admin.proto index b3c23af..8d0398f 100644 --- hbase-protocol/src/main/protobuf/Admin.proto +++ hbase-protocol/src/main/protobuf/Admin.proto @@ -30,12 +30,12 @@ import "WAL.proto"; message GetRegionInfoRequest { required RegionSpecifier region = 1; - optional bool compactionState = 2; + optional bool compaction_state = 2; } message GetRegionInfoResponse { - required RegionInfo regionInfo = 1; - optional CompactionState compactionState = 2; + required RegionInfo region_info = 1; + optional CompactionState compaction_state = 2; enum CompactionState { NONE = 0; @@ -55,28 +55,28 @@ message GetStoreFileRequest { } message GetStoreFileResponse { - repeated string storeFile = 1; + repeated string store_file = 1; } message GetOnlineRegionRequest { } message GetOnlineRegionResponse { - repeated RegionInfo regionInfo = 1; + repeated RegionInfo region_info = 1; } message OpenRegionRequest { - repeated RegionOpenInfo openInfo = 1; + repeated RegionOpenInfo open_info = 1; message RegionOpenInfo { required RegionInfo region = 1; - optional uint32 versionOfOfflineNode = 2; - repeated ServerName favoredNodes = 3; + optional uint32 version_of_offline_node = 2; + repeated ServerName favored_nodes = 3; } } message OpenRegionResponse { - repeated RegionOpeningState openingState = 1; + repeated RegionOpeningState opening_state = 1; enum RegionOpeningState { OPENED = 0; @@ -91,9 +91,9 @@ message OpenRegionResponse { */ message CloseRegionRequest { required RegionSpecifier region = 1; - optional uint32 versionOfClosingNode = 2; - optional bool transitionInZK = 3 [default = true]; - optional ServerName destinationServer = 4; + optional uint32 version_of_closing_node = 2; + optional bool transition_in_ZK = 3 [default = true]; + optional ServerName destination_server = 4; } message CloseRegionResponse { @@ -107,11 +107,11 @@ message CloseRegionResponse { */ message FlushRegionRequest { required RegionSpecifier region = 1; - optional uint64 ifOlderThanTs = 2; + optional uint64 if_older_than_ts = 2; } message FlushRegionResponse { - required uint64 lastFlushTime = 1; + required uint64 last_flush_time = 1; optional bool flushed = 2; } @@ -124,7 +124,7 @@ message FlushRegionResponse { */ message SplitRegionRequest { required RegionSpecifier region = 1; - optional bytes splitPoint = 2; + optional bytes split_point = 2; } message SplitRegionResponse { @@ -150,8 +150,8 @@ message CompactRegionResponse { * This method currently closes the regions and then merges them */ message MergeRegionsRequest { - required RegionSpecifier regionA = 1; - required RegionSpecifier regionB = 2; + required RegionSpecifier region_a = 1; + required RegionSpecifier region_b = 2; optional bool forcible = 3 [default = false]; } @@ -163,10 +163,10 @@ message WALEntry { required WALKey key = 1; // Following may be null if the KVs/Cells are carried along the side in a cellblock (See // RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null - // and associatedCellCount has count of Cells associated w/ this WALEntry - repeated bytes keyValueBytes = 2; + // and associated_cell_count has count of Cells associated w/ this WALEntry + repeated bytes key_value_bytes = 2; // If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock. - optional int32 associatedCellCount = 3; + optional int32 associated_cell_count = 3; } /** @@ -186,7 +186,7 @@ message RollWALWriterRequest { message RollWALWriterResponse { // A list of encoded name of regions to flush - repeated bytes regionToFlush = 1; + repeated bytes region_to_flush = 1; } message StopServerRequest { @@ -200,54 +200,54 @@ message GetServerInfoRequest { } message ServerInfo { - required ServerName serverName = 1; - optional uint32 webuiPort = 2; + required ServerName server_name = 1; + optional uint32 webui_port = 2; } message GetServerInfoResponse { - required ServerInfo serverInfo = 1; + required ServerInfo server_info = 1; } service AdminService { - rpc getRegionInfo(GetRegionInfoRequest) + rpc GetRegionInfo(GetRegionInfoRequest) returns(GetRegionInfoResponse); - rpc getStoreFile(GetStoreFileRequest) + rpc GetStoreFile(GetStoreFileRequest) returns(GetStoreFileResponse); - rpc getOnlineRegion(GetOnlineRegionRequest) + rpc GetOnlineRegion(GetOnlineRegionRequest) returns(GetOnlineRegionResponse); - rpc openRegion(OpenRegionRequest) + rpc OpenRegion(OpenRegionRequest) returns(OpenRegionResponse); - rpc closeRegion(CloseRegionRequest) + rpc CloseRegion(CloseRegionRequest) returns(CloseRegionResponse); - rpc flushRegion(FlushRegionRequest) + rpc FlushRegion(FlushRegionRequest) returns(FlushRegionResponse); - rpc splitRegion(SplitRegionRequest) + rpc SplitRegion(SplitRegionRequest) returns(SplitRegionResponse); - rpc compactRegion(CompactRegionRequest) + rpc CompactRegion(CompactRegionRequest) returns(CompactRegionResponse); - - rpc mergeRegions(MergeRegionsRequest) + + rpc MergeRegions(MergeRegionsRequest) returns(MergeRegionsResponse); - rpc replicateWALEntry(ReplicateWALEntryRequest) + rpc ReplicateWALEntry(ReplicateWALEntryRequest) returns(ReplicateWALEntryResponse); - - rpc replay(MultiRequest) - returns(MultiResponse); - rpc rollWALWriter(RollWALWriterRequest) + rpc Replay(MultiRequest) + returns(MultiResponse); + + rpc RollWALWriter(RollWALWriterRequest) returns(RollWALWriterResponse); - rpc getServerInfo(GetServerInfoRequest) + rpc GetServerInfo(GetServerInfoRequest) returns(GetServerInfoResponse); - rpc stopServer(StopServerRequest) + rpc StopServer(StopServerRequest) returns(StopServerResponse); } diff --git hbase-protocol/src/main/protobuf/Aggregate.proto hbase-protocol/src/main/protobuf/Aggregate.proto index 5707e14..23dbf33 100644 --- hbase-protocol/src/main/protobuf/Aggregate.proto +++ hbase-protocol/src/main/protobuf/Aggregate.proto @@ -31,32 +31,32 @@ message AggregateArgument { * (3) any bytes required to construct the ColumnInterpreter object * properly */ - required string interpreterClassName = 1; + required string interpreter_class_name = 1; required Scan scan = 2; - optional bytes interpreterSpecificBytes = 3; + optional bytes interpreter_specific_bytes = 3; } message AggregateResponse { /** * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both firstPart and secondPart - * have defined values (and the secondPart is not present in the response + * or a simple object. When it is a Pair both first_part and second_part + * have defined values (and the second_part is not present in the response * when the response is not a pair). Refer to the AggregateImplementation * class for an overview of the AggregateResponse object constructions. */ - repeated bytes firstPart = 1; - optional bytes secondPart = 2; + repeated bytes first_part = 1; + optional bytes second_part = 2; } /** Refer to the AggregateImplementation class for an overview of the * AggregateService method implementations and their functionality. */ service AggregateService { - rpc getMax (AggregateArgument) returns (AggregateResponse); - rpc getMin (AggregateArgument) returns (AggregateResponse); - rpc getSum (AggregateArgument) returns (AggregateResponse); - rpc getRowNum (AggregateArgument) returns (AggregateResponse); - rpc getAvg (AggregateArgument) returns (AggregateResponse); - rpc getStd (AggregateArgument) returns (AggregateResponse); - rpc getMedian (AggregateArgument) returns (AggregateResponse); -} \ No newline at end of file + rpc GetMax (AggregateArgument) returns (AggregateResponse); + rpc GetMin (AggregateArgument) returns (AggregateResponse); + rpc GetSum (AggregateArgument) returns (AggregateResponse); + rpc GetRowNum (AggregateArgument) returns (AggregateResponse); + rpc GetAvg (AggregateArgument) returns (AggregateResponse); + rpc GetStd (AggregateArgument) returns (AggregateResponse); + rpc GetMedian (AggregateArgument) returns (AggregateResponse); +} diff --git hbase-protocol/src/main/protobuf/Authentication.proto hbase-protocol/src/main/protobuf/Authentication.proto index dbaab28..1fb38bb 100644 --- hbase-protocol/src/main/protobuf/Authentication.proto +++ hbase-protocol/src/main/protobuf/Authentication.proto @@ -24,7 +24,7 @@ option optimize_for = SPEED; message AuthenticationKey { required int32 id = 1; - required int64 expirationDate = 2; + required int64 expiration_date = 2; required bytes key = 3; } @@ -35,10 +35,10 @@ message TokenIdentifier { } required Kind kind = 1; required bytes username = 2; - required int32 keyId = 3; - optional int64 issueDate = 4; - optional int64 expirationDate = 5; - optional int64 sequenceNumber = 6; + required int32 key_id = 3; + optional int64 issue_date = 4; + optional int64 expiration_date = 5; + optional int64 sequence_number = 6; } @@ -67,15 +67,15 @@ message WhoAmIRequest { message WhoAmIResponse { optional string username = 1; - optional string authMethod = 2; + optional string auth_method = 2; } // RPC service service AuthenticationService { - rpc getAuthenticationToken(TokenRequest) + rpc GetAuthenticationToken(TokenRequest) returns (TokenResponse); - rpc whoami(WhoAmIRequest) + rpc WhoAmI(WhoAmIRequest) returns (WhoAmIResponse); -} \ No newline at end of file +} diff --git hbase-protocol/src/main/protobuf/Cell.proto hbase-protocol/src/main/protobuf/Cell.proto index 85f6eeb..29e1874 100644 --- hbase-protocol/src/main/protobuf/Cell.proto +++ hbase-protocol/src/main/protobuf/Cell.proto @@ -46,7 +46,7 @@ message Cell { optional bytes family = 2; optional bytes qualifier = 3; optional uint64 timestamp = 4; - optional CellType cellType = 5; + optional CellType cell_type = 5; optional bytes value = 6; } @@ -59,6 +59,6 @@ message KeyValue { required bytes family = 2; required bytes qualifier = 3; optional uint64 timestamp = 4; - optional CellType keyType = 5; + optional CellType key_type = 5; optional bytes value = 6; } diff --git hbase-protocol/src/main/protobuf/Client.proto hbase-protocol/src/main/protobuf/Client.proto index e89d184..2d0892f 100644 --- hbase-protocol/src/main/protobuf/Client.proto +++ hbase-protocol/src/main/protobuf/Client.proto @@ -45,11 +45,11 @@ message Get { repeated Column column = 2; repeated NameBytesPair attribute = 3; optional Filter filter = 4; - optional TimeRange timeRange = 5; - optional uint32 maxVersions = 6 [default = 1]; - optional bool cacheBlocks = 7 [default = true]; - optional uint32 storeLimit = 8; - optional uint32 storeOffset = 9; + optional TimeRange time_range = 5; + optional uint32 max_versions = 6 [default = 1]; + optional bool cache_blocks = 7 [default = true]; + optional uint32 store_limit = 8; + optional uint32 store_offset = 9; } message Result { @@ -60,18 +60,18 @@ message Result { // not part of this protobuf message; they are passed alongside // and then this Message is just a placeholder with metadata. // The count is needed to know how many to peel off the block of Cells as - // ours. NOTE: This is different from the pb managed cellCount of the + // ours. NOTE: This is different from the pb managed cell_count of the // 'cell' field above which is non-null when the cells are pb'd. - optional int32 associatedCellCount = 2; + optional int32 associated_cell_count = 2; } /** * The get request. Perform a single Get operation. - * Unless existenceOnly is specified, return all the requested data + * Unless existence_only is specified, return all the requested data * for the row that matches exactly, or the one that immediately - * precedes it if closestRowBefore is specified. + * precedes it if closest_row_before is specified. * - * If existenceOnly is set, only the existence will be returned. + * If existence_only is set, only the existence will be returned. */ message GetRequest { required RegionSpecifier region = 1; @@ -79,12 +79,12 @@ message GetRequest { // If the row to get doesn't exist, return the // closest row before. - optional bool closestRowBefore = 3; + optional bool closest_row_before = 3; // The result isn't asked for, just check for - // the existence. If closestRowBefore specified, + // the existence. If closest_row_before specified, // this will be ignored - optional bool existenceOnly = 4; + optional bool existence_only = 4; } message MultiGetRequest { @@ -93,12 +93,12 @@ message MultiGetRequest { // If the row to get doesn't exist, return the // closest row before. - optional bool closestRowBefore = 3; + optional bool closest_row_before = 3; // The result isn't asked for, just check for - // the existence. If closestRowBefore specified, + // the existence. If closest_row_before specified, // this will be ignored - optional bool existenceOnly = 4; + optional bool existence_only = 4; } message GetResponse { @@ -125,7 +125,7 @@ message Condition { required bytes row = 1; required bytes family = 2; required bytes qualifier = 3; - required CompareType compareType = 4; + required CompareType compare_type = 4; required Comparator comparator = 5; } @@ -139,22 +139,22 @@ message Condition { */ message MutationProto { optional bytes row = 1; - optional MutationType mutateType = 2; - repeated ColumnValue columnValue = 3; + optional MutationType mutate_type = 2; + repeated ColumnValue column_value = 3; optional uint64 timestamp = 4; repeated NameBytesPair attribute = 5; optional Durability durability = 6 [default = USE_DEFAULT]; // For some mutations, a result may be returned, in which case, // time range can be specified for potential performance gain - optional TimeRange timeRange = 7; + optional TimeRange time_range = 7; // The below count is set when the associated cells are NOT // part of this protobuf message; they are passed alongside // and then this Message is a placeholder with metadata. The // count is needed to know how many to peel off the block of Cells as - // ours. NOTE: This is different from the pb managed cellCount of the + // ours. NOTE: This is different from the pb managed cell_count of the // 'cell' field above which is non-null when the cells are pb'd. - optional int32 associatedCellCount = 8; + optional int32 associated_cell_count = 8; enum Durability { USE_DEFAULT = 0; @@ -180,13 +180,13 @@ message MutationProto { message ColumnValue { required bytes family = 1; - repeated QualifierValue qualifierValue = 2; + repeated QualifierValue qualifier_value = 2; message QualifierValue { optional bytes qualifier = 1; optional bytes value = 2; optional uint64 timestamp = 3; - optional DeleteType deleteType = 4; + optional DeleteType delete_type = 4; } } } @@ -225,18 +225,18 @@ message MutateResponse { message Scan { repeated Column column = 1; repeated NameBytesPair attribute = 2; - optional bytes startRow = 3; - optional bytes stopRow = 4; + optional bytes start_row = 3; + optional bytes stop_row = 4; optional Filter filter = 5; - optional TimeRange timeRange = 6; - optional uint32 maxVersions = 7 [default = 1]; - optional bool cacheBlocks = 8 [default = true]; - optional uint32 batchSize = 9; - optional uint64 maxResultSize = 10; - optional uint32 storeLimit = 11; - optional uint32 storeOffset = 12; - optional bool loadColumnFamiliesOnDemand = 13; /* DO NOT add defaults to loadColumnFamiliesOnDemand. */ - optional uint32 cachingCount = 14; + optional TimeRange time_range = 6; + optional uint32 max_versions = 7 [default = 1]; + optional bool cache_blocks = 8 [default = true]; + optional uint32 batch_size = 9; + optional uint64 max_result_size = 10; + optional uint32 store_limit = 11; + optional uint32 store_offset = 12; + optional bool load_column_families_on_demand = 13; /* DO NOT add defaults to load_column_families_on_demand. */ + optional uint32 caching_count = 14; optional bool prefetching = 15; } @@ -254,25 +254,25 @@ message Scan { message ScanRequest { optional RegionSpecifier region = 1; optional Scan scan = 2; - optional uint64 scannerId = 3; - optional uint32 numberOfRows = 4; - optional bool closeScanner = 5; - optional uint64 nextCallSeq = 6; + optional uint64 scanner_id = 3; + optional uint32 number_of_rows = 4; + optional bool close_scanner = 5; + optional uint64 next_call_seq = 6; } /** - * The scan response. If there are no more results, moreResults will + * The scan response. If there are no more results, more_results will * be false. If it is not specified, it means there are more. */ message ScanResponse { - optional ResultCellMeta resultCellMeta = 1; - optional uint64 scannerId = 2; - optional bool moreResults = 3; + optional ResultCellMeta result_cell_meta = 1; + optional uint64 scanner_id = 2; + optional bool more_results = 3; optional uint32 ttl = 4; } message ResultCellMeta { - repeated uint32 cellsLength = 1; + repeated uint32 cells_length = 1; } /** @@ -281,8 +281,8 @@ message ResultCellMeta { */ message BulkLoadHFileRequest { required RegionSpecifier region = 1; - repeated FamilyPath familyPath = 2; - optional bool assignSeqNum = 3; + repeated FamilyPath family_path = 2; + optional bool assign_seq_num = 3; message FamilyPath { required bytes family = 1; @@ -296,8 +296,8 @@ message BulkLoadHFileResponse { message CoprocessorServiceCall { required bytes row = 1; - required string serviceName = 2; - required string methodName = 3; + required string service_name = 2; + required string method_name = 3; required bytes request = 4; } @@ -352,24 +352,24 @@ message MultiResponse { service ClientService { - rpc get(GetRequest) + rpc Get(GetRequest) returns(GetResponse); - rpc multiGet(MultiGetRequest) + rpc MultiGet(MultiGetRequest) returns(MultiGetResponse); - rpc mutate(MutateRequest) + rpc Mutate(MutateRequest) returns(MutateResponse); - rpc scan(ScanRequest) + rpc Scan(ScanRequest) returns(ScanResponse); - rpc bulkLoadHFile(BulkLoadHFileRequest) + rpc BulkLoadHFile(BulkLoadHFileRequest) returns(BulkLoadHFileResponse); - rpc execService(CoprocessorServiceRequest) + rpc ExecService(CoprocessorServiceRequest) returns(CoprocessorServiceResponse); - rpc multi(MultiRequest) + rpc Multi(MultiRequest) returns(MultiResponse); } diff --git hbase-protocol/src/main/protobuf/ClusterId.proto hbase-protocol/src/main/protobuf/ClusterId.proto index 80b7d0f..b3116f7 100644 --- hbase-protocol/src/main/protobuf/ClusterId.proto +++ hbase-protocol/src/main/protobuf/ClusterId.proto @@ -29,5 +29,5 @@ option optimize_for = SPEED; */ message ClusterId { // This is the cluster id, a uuid as a String - required string clusterId = 1; + required string cluster_id = 1; } diff --git hbase-protocol/src/main/protobuf/ClusterStatus.proto hbase-protocol/src/main/protobuf/ClusterStatus.proto index 9009144..bdaef39 100644 --- hbase-protocol/src/main/protobuf/ClusterStatus.proto +++ hbase-protocol/src/main/protobuf/ClusterStatus.proto @@ -28,7 +28,7 @@ import "ClusterId.proto"; import "FS.proto"; message RegionState { - required RegionInfo regionInfo = 1; + required RegionInfo region_info = 1; required State state = 2; optional uint64 stamp = 3; enum State { @@ -50,12 +50,12 @@ message RegionState { message RegionInTransition { required RegionSpecifier spec = 1; - required RegionState regionState = 2; + required RegionState region_state = 2; } message RegionLoad { /** the region specifier */ - required RegionSpecifier regionSpecifier = 1; + required RegionSpecifier region_specifier = 1; /** the number of stores for the region */ optional uint32 stores = 2; @@ -64,65 +64,65 @@ message RegionLoad { optional uint32 storefiles = 3; /** the total size of the store files for the region, uncompressed, in MB */ - optional uint32 storeUncompressedSizeMB = 4; + optional uint32 store_uncompressed_size_MB = 4; /** the current total size of the store files for the region, in MB */ - optional uint32 storefileSizeMB = 5; + optional uint32 storefile_size_MB = 5; /** the current size of the memstore for the region, in MB */ - optional uint32 memstoreSizeMB = 6; + optional uint32 memstore_size_MB = 6; /** * The current total size of root-level store file indexes for the region, * in MB. The same as {@link #rootIndexSizeKB} but in MB. */ - optional uint32 storefileIndexSizeMB = 7; + optional uint32 storefile_index_size_MB = 7; /** the current total read requests made to region */ - optional uint64 readRequestsCount = 8; + optional uint64 read_requests_count = 8; /** the current total write requests made to region */ - optional uint64 writeRequestsCount = 9; + optional uint64 write_requests_count = 9; /** the total compacting key values in currently running compaction */ - optional uint64 totalCompactingKVs = 10; + optional uint64 total_compacting_KVs = 10; /** the completed count of key values in currently running compaction */ - optional uint64 currentCompactedKVs = 11; + optional uint64 current_compacted_KVs = 11; /** The current total size of root-level indexes for the region, in KB. */ - optional uint32 rootIndexSizeKB = 12; + optional uint32 root_index_size_KB = 12; /** The total size of all index blocks, not just the root level, in KB. */ - optional uint32 totalStaticIndexSizeKB = 13; + optional uint32 total_static_index_size_KB = 13; /** * The total size of all Bloom filter blocks, not just loaded into the * block cache, in KB. */ - optional uint32 totalStaticBloomSizeKB = 14; + optional uint32 total_static_bloom_size_KB = 14; /** the most recent sequence Id from cache flush */ - optional uint64 completeSequenceId = 15; + optional uint64 complete_sequence_id = 15; } /* Server-level protobufs */ message ServerLoad { /** Number of requests since last report. */ - optional uint32 numberOfRequests = 1; + optional uint32 number_of_requests = 1; /** Total Number of requests from the start of the region server. */ - optional uint32 totalNumberOfRequests = 2; + optional uint32 total_number_of_requests = 2; /** the amount of used heap, in MB. */ - optional uint32 usedHeapMB = 3; + optional uint32 used_heap_MB = 3; /** the maximum allowable size of the heap, in MB. */ - optional uint32 maxHeapMB = 4; + optional uint32 max_heap_MB = 4; /** Information on the load of individual regions. */ - repeated RegionLoad regionLoads = 5; + repeated RegionLoad region_loads = 5; /** * Regionserver-level coprocessors, e.g., WALObserver implementations. @@ -132,39 +132,38 @@ message ServerLoad { repeated Coprocessor coprocessors = 6; /** - * Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests) + * Time when incremental (non-total) counts began being calculated (e.g. number_of_requests) * time is measured as the difference, measured in milliseconds, between the current time * and midnight, January 1, 1970 UTC. */ - optional uint64 reportStartTime = 7; + optional uint64 report_start_time = 7; /** * Time when report was generated. * time is measured as the difference, measured in milliseconds, between the current time * and midnight, January 1, 1970 UTC. */ - optional uint64 reportEndTime = 8; + optional uint64 report_end_time = 8; /** * The port number that this region server is hosing an info server on. */ - optional uint32 infoServerPort = 9; + optional uint32 info_server_port = 9; } message LiveServerInfo { required ServerName server = 1; - required ServerLoad serverLoad = 2; + required ServerLoad server_load = 2; } - message ClusterStatus { - optional HBaseVersionFileContent hbaseVersion = 1; - repeated LiveServerInfo liveServers = 2; - repeated ServerName deadServers = 3; - repeated RegionInTransition regionsInTransition = 4; - optional ClusterId clusterId = 5; - repeated Coprocessor masterCoprocessors = 6; + optional HBaseVersionFileContent hbase_version = 1; + repeated LiveServerInfo live_servers = 2; + repeated ServerName dead_servers = 3; + repeated RegionInTransition regions_in_transition = 4; + optional ClusterId cluster_id = 5; + repeated Coprocessor master_coprocessors = 6; optional ServerName master = 7; - repeated ServerName backupMasters = 8; - optional bool balancerOn = 9; + repeated ServerName backup_masters = 8; + optional bool balancer_on = 9; } diff --git hbase-protocol/src/main/protobuf/Comparator.proto hbase-protocol/src/main/protobuf/Comparator.proto index d4c1682..f6daf81 100644 --- hbase-protocol/src/main/protobuf/Comparator.proto +++ hbase-protocol/src/main/protobuf/Comparator.proto @@ -28,7 +28,7 @@ option optimize_for = SPEED; message Comparator { required string name = 1; - optional bytes serializedComparator = 2; + optional bytes serialized_comparator = 2; } message ByteArrayComparable { @@ -45,7 +45,7 @@ message BinaryPrefixComparator { message BitComparator { required ByteArrayComparable comparable = 1; - required BitwiseOp bitwiseOp = 2; + required BitwiseOp bitwise_op = 2; enum BitwiseOp { AND = 1; @@ -59,7 +59,7 @@ message NullComparator { message RegexStringComparator { required string pattern = 1; - required int32 patternFlags = 2; + required int32 pattern_flags = 2; required string charset = 3; } diff --git hbase-protocol/src/main/protobuf/ErrorHandling.proto hbase-protocol/src/main/protobuf/ErrorHandling.proto index e3708bc..1c1488c 100644 --- hbase-protocol/src/main/protobuf/ErrorHandling.proto +++ hbase-protocol/src/main/protobuf/ErrorHandling.proto @@ -28,10 +28,10 @@ option optimize_for = SPEED; * so we can serialize exceptions. */ message StackTraceElementMessage { - optional string declaringClass = 1; - optional string methodName = 2; - optional string fileName = 3; - optional int32 lineNumber = 4; + optional string declaring_class = 1; + optional string method_name = 2; + optional string file_name = 3; + optional int32 line_number = 4; } /** @@ -41,9 +41,9 @@ message StackTraceElementMessage { * (which should be another protobuffed class). */ message GenericExceptionMessage { - optional string className = 1; + optional string class_name = 1; optional string message = 2; - optional bytes errorInfo = 3; + optional bytes error_info = 3; repeated StackTraceElementMessage trace = 4; } @@ -53,6 +53,5 @@ message GenericExceptionMessage { */ message ForeignExceptionMessage { optional string source = 1; - optional GenericExceptionMessage genericException = 2; - -} \ No newline at end of file + optional GenericExceptionMessage generic_exception = 2; +} diff --git hbase-protocol/src/main/protobuf/Filter.proto hbase-protocol/src/main/protobuf/Filter.proto index 618ef30..d376218 100644 --- hbase-protocol/src/main/protobuf/Filter.proto +++ hbase-protocol/src/main/protobuf/Filter.proto @@ -29,7 +29,7 @@ import "Comparator.proto"; message Filter { required string name = 1; - optional bytes serializedFilter = 2; + optional bytes serialized_filter = 2; } message ColumnCountGetFilter { @@ -39,7 +39,7 @@ message ColumnCountGetFilter { message ColumnPaginationFilter { required int32 limit = 1; optional int32 offset = 2; - optional bytes columnOffset = 3; + optional bytes column_offset = 3; } message ColumnPrefixFilter { @@ -47,26 +47,26 @@ message ColumnPrefixFilter { } message ColumnRangeFilter { - optional bytes minColumn = 1; - optional bool minColumnInclusive = 2; - optional bytes maxColumn = 3; - optional bool maxColumnInclusive = 4; + optional bytes min_column = 1; + optional bool min_column_inclusive = 2; + optional bytes max_column = 3; + optional bool max_column_inclusive = 4; } message CompareFilter { - required CompareType compareOp = 1; + required CompareType compare_op = 1; optional Comparator comparator = 2; } message DependentColumnFilter { - required CompareFilter compareFilter = 1; - optional bytes columnFamily = 2; - optional bytes columnQualifier = 3; - optional bool dropDependentColumn = 4; + required CompareFilter compare_filter = 1; + optional bytes column_family = 2; + optional bytes column_qualifier = 3; + optional bool drop_dependent_column = 4; } message FamilyFilter { - required CompareFilter compareFilter = 1; + required CompareFilter compare_filter = 1; } message FilterList { @@ -91,23 +91,23 @@ message FirstKeyValueMatchingQualifiersFilter { } message FuzzyRowFilter { - repeated BytesBytesPair fuzzyKeysData = 1; + repeated BytesBytesPair fuzzy_keys_data = 1; } message InclusiveStopFilter { - optional bytes stopRowKey = 1; + optional bytes stop_row_key = 1; } message KeyOnlyFilter { - required bool lenAsVal = 1; + required bool len_as_val = 1; } message MultipleColumnPrefixFilter { - repeated bytes sortedPrefixes = 1; + repeated bytes sorted_prefixes = 1; } message PageFilter { - required int64 pageSize = 1; + required int64 page_size = 1; } message PrefixFilter { @@ -115,7 +115,7 @@ message PrefixFilter { } message QualifierFilter { - required CompareFilter compareFilter = 1; + required CompareFilter compare_filter = 1; } message RandomRowFilter { @@ -123,20 +123,20 @@ message RandomRowFilter { } message RowFilter { - required CompareFilter compareFilter = 1; + required CompareFilter compare_filter = 1; } message SingleColumnValueExcludeFilter { - required SingleColumnValueFilter singleColumnValueFilter = 1; + required SingleColumnValueFilter single_column_value_filter = 1; } message SingleColumnValueFilter { - optional bytes columnFamily = 1; - optional bytes columnQualifier = 2; - required CompareType compareOp = 3; + optional bytes column_family = 1; + optional bytes column_qualifier = 2; + required CompareType compare_op = 3; required Comparator comparator = 4; - optional bool filterIfMissing = 5; - optional bool latestVersionOnly = 6; + optional bool filter_if_missing = 5; + optional bool latest_version_only = 6; } message SkipFilter { @@ -148,7 +148,7 @@ message TimestampsFilter { } message ValueFilter { - required CompareFilter compareFilter = 1; + required CompareFilter compare_filter = 1; } message WhileMatchFilter { diff --git hbase-protocol/src/main/protobuf/HFile.proto hbase-protocol/src/main/protobuf/HFile.proto index b843f5c..17ae5d9 100644 --- hbase-protocol/src/main/protobuf/HFile.proto +++ hbase-protocol/src/main/protobuf/HFile.proto @@ -25,21 +25,21 @@ import "hbase.proto"; // Map of name/values message FileInfoProto { - repeated BytesBytesPair mapEntry = 1; + repeated BytesBytesPair map_entry = 1; } // HFile file trailer message FileTrailerProto { - optional uint64 fileInfoOffset = 1; - optional uint64 loadOnOpenDataOffset = 2; - optional uint64 uncompressedDataIndexSize = 3; - optional uint64 totalUncompressedBytes = 4; - optional uint32 dataIndexCount = 5; - optional uint32 metaIndexCount = 6; - optional uint64 entryCount = 7; - optional uint32 numDataIndexLevels = 8; - optional uint64 firstDataBlockOffset = 9; - optional uint64 lastDataBlockOffset = 10; - optional string comparatorClassName = 11; - optional uint32 compressionCodec = 12; + optional uint64 file_info_offset = 1; + optional uint64 load_on_open_data_offset = 2; + optional uint64 uncompressed_data_index_size = 3; + optional uint64 total_uncompressed_bytes = 4; + optional uint32 data_index_count = 5; + optional uint32 meta_index_count = 6; + optional uint64 entry_count = 7; + optional uint32 num_data_index_levels = 8; + optional uint64 first_data_block_offset = 9; + optional uint64 last_data_block_offset = 10; + optional string comparator_class_name = 11; + optional uint32 compression_codec = 12; } diff --git hbase-protocol/src/main/protobuf/LoadBalancer.proto hbase-protocol/src/main/protobuf/LoadBalancer.proto index df1c049..431b087 100644 --- hbase-protocol/src/main/protobuf/LoadBalancer.proto +++ hbase-protocol/src/main/protobuf/LoadBalancer.proto @@ -24,5 +24,5 @@ option java_generate_equals_and_hash = true; option optimize_for = SPEED; message LoadBalancerState { - optional bool balancerOn = 1; + optional bool balancer_on = 1; } diff --git hbase-protocol/src/main/protobuf/Master.proto hbase-protocol/src/main/protobuf/Master.proto index e2ff35b..afe7b8a 100644 --- hbase-protocol/src/main/protobuf/Master.proto +++ hbase-protocol/src/main/protobuf/Master.proto @@ -28,11 +28,11 @@ message IsMasterRunningRequest { } message IsMasterRunningResponse { - required bool isMasterRunning = 1; + required bool is_master_running = 1; } service MasterService { /** return true if master is available */ - rpc isMasterRunning(IsMasterRunningRequest) + rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); } diff --git hbase-protocol/src/main/protobuf/MasterAdmin.proto hbase-protocol/src/main/protobuf/MasterAdmin.proto index 5339e36..d9e41c0 100644 --- hbase-protocol/src/main/protobuf/MasterAdmin.proto +++ hbase-protocol/src/main/protobuf/MasterAdmin.proto @@ -32,24 +32,24 @@ import "Client.proto"; /* Column-level protobufs */ message AddColumnRequest { - required bytes tableName = 1; - required ColumnFamilySchema columnFamilies = 2; + required bytes table_name = 1; + required ColumnFamilySchema column_families = 2; } message AddColumnResponse { } message DeleteColumnRequest { - required bytes tableName = 1; - required bytes columnName = 2; + required bytes table_name = 1; + required bytes column_name = 2; } message DeleteColumnResponse { } message ModifyColumnRequest { - required bytes tableName = 1; - required ColumnFamilySchema columnFamilies = 2; + required bytes table_name = 1; + required ColumnFamilySchema column_families = 2; } message ModifyColumnResponse { @@ -59,7 +59,7 @@ message ModifyColumnResponse { message MoveRegionRequest { required RegionSpecifier region = 1; - optional ServerName destServerName = 2; + optional ServerName dest_server_name = 2; } message MoveRegionResponse { @@ -69,8 +69,8 @@ message MoveRegionResponse { * Dispatch merging the specified regions. */ message DispatchMergingRegionsRequest { - required RegionSpecifier regionA = 1; - required RegionSpecifier regionB = 2; + required RegionSpecifier region_a = 1; + required RegionSpecifier region_b = 2; optional bool forcible = 3 [default = false]; } @@ -102,37 +102,37 @@ message OfflineRegionResponse { /* Table-level protobufs */ message CreateTableRequest { - required TableSchema tableSchema = 1; - repeated bytes splitKeys = 2; + required TableSchema table_schema = 1; + repeated bytes split_keys = 2; } message CreateTableResponse { } message DeleteTableRequest { - required bytes tableName = 1; + required bytes table_name = 1; } message DeleteTableResponse { } message EnableTableRequest { - required bytes tableName = 1; + required bytes table_name = 1; } message EnableTableResponse { } message DisableTableRequest { - required bytes tableName = 1; + required bytes table_name = 1; } message DisableTableResponse { } message ModifyTableRequest { - required bytes tableName = 1; - required TableSchema tableSchema = 2; + required bytes table_name = 1; + required TableSchema table_schema = 2; } message ModifyTableResponse { @@ -157,7 +157,7 @@ message BalanceRequest { } message BalanceResponse { - required bool balancerRan = 1; + required bool balancer_ran = 1; } message SetBalancerRunningRequest { @@ -166,14 +166,14 @@ message SetBalancerRunningRequest { } message SetBalancerRunningResponse { - optional bool prevBalanceValue = 1; + optional bool prev_balance_value = 1; } message CatalogScanRequest { } message CatalogScanResponse { - optional int32 scanResult = 1; + optional int32 scan_result = 1; } message EnableCatalogJanitorRequest { @@ -181,7 +181,7 @@ message EnableCatalogJanitorRequest { } message EnableCatalogJanitorResponse { - optional bool prevValue = 1; + optional bool prev_value = 1; } message IsCatalogJanitorEnabledRequest { @@ -196,7 +196,7 @@ message TakeSnapshotRequest{ } message TakeSnapshotResponse{ - required int64 expectedTimeout = 1; + required int64 expected_timeout = 1; } message ListSnapshotRequest{ @@ -242,36 +242,36 @@ message IsRestoreSnapshotDoneResponse { service MasterAdminService { /** Adds a column to the specified table. */ - rpc addColumn(AddColumnRequest) + rpc AddColumn(AddColumnRequest) returns(AddColumnResponse); /** Deletes a column from the specified table. Table must be disabled. */ - rpc deleteColumn(DeleteColumnRequest) + rpc DeleteColumn(DeleteColumnRequest) returns(DeleteColumnResponse); /** Modifies an existing column on the specified table. */ - rpc modifyColumn(ModifyColumnRequest) + rpc ModifyColumn(ModifyColumnRequest) returns(ModifyColumnResponse); /** Move the region region to the destination server. */ - rpc moveRegion(MoveRegionRequest) + rpc MoveRegion(MoveRegionRequest) returns(MoveRegionResponse); /** Master dispatch merging the regions */ - rpc dispatchMergingRegions(DispatchMergingRegionsRequest) + rpc DispatchMergingRegions(DispatchMergingRegionsRequest) returns(DispatchMergingRegionsResponse); /** Assign a region to a server chosen at random. */ - rpc assignRegion(AssignRegionRequest) + rpc AssignRegion(AssignRegionRequest) returns(AssignRegionResponse); /** * Unassign a region from current hosting regionserver. Region will then be * assigned to a regionserver chosen at random. Region could be reassigned - * back to the same server. Use moveRegion if you want + * back to the same server. Use MoveRegion if you want * to control the region movement. */ - rpc unassignRegion(UnassignRegionRequest) + rpc UnassignRegion(UnassignRegionRequest) returns(UnassignRegionResponse); /** @@ -280,35 +280,35 @@ service MasterAdminService { * automatically reassign the region as in unassign. This is a special * method, and should only be used by experts or hbck. */ - rpc offlineRegion(OfflineRegionRequest) + rpc OfflineRegion(OfflineRegionRequest) returns(OfflineRegionResponse); /** Deletes a table */ - rpc deleteTable(DeleteTableRequest) + rpc DeleteTable(DeleteTableRequest) returns(DeleteTableResponse); /** Puts the table on-line (only needed if table has been previously taken offline) */ - rpc enableTable(EnableTableRequest) + rpc EnableTable(EnableTableRequest) returns(EnableTableResponse); /** Take table offline */ - rpc disableTable(DisableTableRequest) + rpc DisableTable(DisableTableRequest) returns(DisableTableResponse); /** Modify a table's metadata */ - rpc modifyTable(ModifyTableRequest) + rpc ModifyTable(ModifyTableRequest) returns(ModifyTableResponse); /** Creates a new table asynchronously */ - rpc createTable(CreateTableRequest) + rpc CreateTable(CreateTableRequest) returns(CreateTableResponse); /** Shutdown an HBase cluster. */ - rpc shutdown(ShutdownRequest) + rpc Shutdown(ShutdownRequest) returns(ShutdownResponse); /** Stop HBase Master only. Does not shutdown the cluster. */ - rpc stopMaster(StopMasterRequest) + rpc StopMaster(StopMasterRequest) returns(StopMasterResponse); /** @@ -316,72 +316,72 @@ service MasterAdminService { * go ahead and do the reassignments. Can NOT run for various reasons. * Check logs. */ - rpc balance(BalanceRequest) + rpc Balance(BalanceRequest) returns(BalanceResponse); /** * Turn the load balancer on or off. * If synchronous is true, it waits until current balance() call, if outstanding, to return. */ - rpc setBalancerRunning(SetBalancerRunningRequest) + rpc SetBalancerRunning(SetBalancerRunningRequest) returns(SetBalancerRunningResponse); /** Get a run of the catalog janitor */ - rpc runCatalogScan(CatalogScanRequest) + rpc RunCatalogScan(CatalogScanRequest) returns(CatalogScanResponse); /** * Enable the catalog janitor on or off. */ - rpc enableCatalogJanitor(EnableCatalogJanitorRequest) + rpc EnableCatalogJanitor(EnableCatalogJanitorRequest) returns(EnableCatalogJanitorResponse); /** * Query whether the catalog janitor is enabled. */ - rpc isCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest) + rpc IsCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest) returns(IsCatalogJanitorEnabledResponse); /** * Call a master coprocessor endpoint */ - rpc execMasterService(CoprocessorServiceRequest) + rpc ExecMasterService(CoprocessorServiceRequest) returns(CoprocessorServiceResponse); /** * Create a snapshot for the given table. * @param snapshot description of the snapshot to take */ - rpc snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse); + rpc Snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse); /** * List completed snapshots. * Returns a list of snapshot descriptors for completed snapshots */ - rpc getCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse); + rpc GetCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse); /** * Delete an existing snapshot. This method can also be used to clean up an aborted snapshot. * @param snapshotName snapshot to delete */ - rpc deleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse); + rpc DeleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse); /** * Determine if the snapshot is done yet. */ - rpc isSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse); + rpc IsSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse); /** * Restore a snapshot * @param snapshot description of the snapshot to restore */ - rpc restoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse); + rpc RestoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse); /** * Determine if the snapshot restore is done yet. */ - rpc isRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse); + rpc IsRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse); /** return true if master is available */ - rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); + rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); } diff --git hbase-protocol/src/main/protobuf/MasterMonitor.proto hbase-protocol/src/main/protobuf/MasterMonitor.proto index a6054cf..44e42c5 100644 --- hbase-protocol/src/main/protobuf/MasterMonitor.proto +++ hbase-protocol/src/main/protobuf/MasterMonitor.proto @@ -29,42 +29,42 @@ import "hbase.proto"; import "ClusterStatus.proto"; message GetSchemaAlterStatusRequest { - required bytes tableName = 1; + required bytes table_name = 1; } message GetSchemaAlterStatusResponse { - optional uint32 yetToUpdateRegions = 1; - optional uint32 totalRegions = 2; + optional uint32 yet_to_update_regions = 1; + optional uint32 total_regions = 2; } message GetTableDescriptorsRequest { - repeated string tableNames = 1; + repeated string table_names = 1; } message GetTableDescriptorsResponse { - repeated TableSchema tableSchema = 1; + repeated TableSchema table_schema = 1; } message GetClusterStatusRequest { } message GetClusterStatusResponse { - required ClusterStatus clusterStatus = 1; + required ClusterStatus cluster_status = 1; } service MasterMonitorService { /** Used by the client to get the number of regions that have received the updated schema */ - rpc getSchemaAlterStatus(GetSchemaAlterStatusRequest) + rpc GetSchemaAlterStatus(GetSchemaAlterStatusRequest) returns(GetSchemaAlterStatusResponse); /** Get list of TableDescriptors for requested tables. */ - rpc getTableDescriptors(GetTableDescriptorsRequest) + rpc GetTableDescriptors(GetTableDescriptorsRequest) returns(GetTableDescriptorsResponse); /** Return cluster status. */ - rpc getClusterStatus(GetClusterStatusRequest) + rpc GetClusterStatus(GetClusterStatusRequest) returns(GetClusterStatusResponse); /** return true if master is available */ - rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); + rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); } diff --git hbase-protocol/src/main/protobuf/MultiRowMutation.proto hbase-protocol/src/main/protobuf/MultiRowMutation.proto index 02b038b..31469a8 100644 --- hbase-protocol/src/main/protobuf/MultiRowMutation.proto +++ hbase-protocol/src/main/protobuf/MultiRowMutation.proto @@ -23,13 +23,13 @@ option java_generic_services = true; option optimize_for = SPEED; message MultiMutateRequest { - repeated MutationProto mutationRequest = 1; + repeated MutationProto mutation_request = 1; } message MultiMutateResponse { } service MultiRowMutationService { - rpc mutateRows(MultiMutateRequest) + rpc MutateRows(MultiMutateRequest) returns(MultiMutateResponse); } diff --git hbase-protocol/src/main/protobuf/RPC.proto hbase-protocol/src/main/protobuf/RPC.proto index 7c65e0f..080c690 100644 --- hbase-protocol/src/main/protobuf/RPC.proto +++ hbase-protocol/src/main/protobuf/RPC.proto @@ -72,20 +72,20 @@ option optimize_for = SPEED; // User Information proto. Included in ConnectionHeader on connection setup message UserInformation { - required string effectiveUser = 1; - optional string realUser = 2; + required string effective_user = 1; + optional string real_user = 2; } // This is sent on connection setup after the connection preamble is sent. message ConnectionHeader { - optional UserInformation userInfo = 1; - optional string serviceName = 2; + optional UserInformation user_info = 1; + optional string service_name = 2; // Cell block codec we will use sending over optional cell blocks. Server throws exception // if cannot deal. - optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; + optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; // Compressor we will use if cell block is compressed. Server will throw exception if not supported. // Class must implement hadoop's CompressionCodec Interface - optional string cellBlockCompressorClass = 4; + optional string cell_block_compressor_class = 4; } // Optional Cell block Message. Included in client RequestHeader @@ -98,34 +98,34 @@ message CellBlockMeta { // the server side exception to the RPC client. message ExceptionResponse { // Class name of the exception thrown from the server - optional string exceptionClassName = 1; + optional string exception_class_name = 1; // Exception stack trace from the server side - optional string stackTrace = 2; + optional string stack_trace = 2; // Optional hostname. Filled in for some exceptions such as region moved // where exception gives clue on where the region may have moved. optional string hostname = 3; optional int32 port = 4; // Set if we are NOT to retry on receipt of this exception - optional bool doNotRetry = 5; + optional bool do_not_retry = 5; } // Header sent making a request. message RequestHeader { - // Monotonically increasing callId to keep track of RPC requests and their response - optional uint32 callId = 1; - optional RPCTInfo traceInfo = 2; - optional string methodName = 3; + // Monotonically increasing call_id to keep track of RPC requests and their response + optional uint32 call_id = 1; + optional RPCTInfo trace_info = 2; + optional string method_name = 3; // If true, then a pb Message param follows. - optional bool requestParam = 4; + optional bool request_param = 4; // If present, then an encoded data block follows. - optional CellBlockMeta cellBlockMeta = 5; + optional CellBlockMeta cell_block_meta = 5; // TODO: Have client specify priority } message ResponseHeader { - optional uint32 callId = 1; + optional uint32 call_id = 1; // If present, then request threw an exception and no response message (else we presume one) optional ExceptionResponse exception = 2; // If present, then an encoded data block follows. - optional CellBlockMeta cellBlockMeta = 3; + optional CellBlockMeta cell_block_meta = 3; } diff --git hbase-protocol/src/main/protobuf/RegionServerStatus.proto hbase-protocol/src/main/protobuf/RegionServerStatus.proto index 767749d..675aab1 100644 --- hbase-protocol/src/main/protobuf/RegionServerStatus.proto +++ hbase-protocol/src/main/protobuf/RegionServerStatus.proto @@ -32,10 +32,10 @@ message RegionServerStartupRequest { required uint32 port = 1; /** This servers' startcode */ - required uint64 serverStartCode = 2; + required uint64 server_start_code = 2; /** Current time of the region server in ms */ - required uint64 serverCurrentTime = 3; + required uint64 server_current_time = 3; } message RegionServerStartupResponse { @@ -44,7 +44,7 @@ message RegionServerStartupResponse { * hbase rootdir, the hostname to use creating the RegionServer ServerName, * etc */ - repeated NameStringPair mapEntries = 1; + repeated NameStringPair map_entries = 1; } message RegionServerReportRequest { @@ -62,7 +62,7 @@ message ReportRSFatalErrorRequest { required ServerName server = 1; /** informative text to expose in the master logs and UI */ - required string errorMessage = 2; + required string error_message = 2; } message ReportRSFatalErrorResponse { @@ -70,33 +70,33 @@ message ReportRSFatalErrorResponse { message GetLastFlushedSequenceIdRequest { /** region name */ - required bytes regionName = 1; + required bytes region_name = 1; } message GetLastFlushedSequenceIdResponse { /** the last HLog sequence id flushed from MemStore to HFile for the region */ - required uint64 lastFlushedSequenceId = 1; + required uint64 last_flushed_sequence_id = 1; } service RegionServerStatusService { /** Called when a region server first starts. */ - rpc regionServerStartup(RegionServerStartupRequest) + rpc RegionServerStartup(RegionServerStartupRequest) returns(RegionServerStartupResponse); /** Called to report the load the RegionServer is under. */ - rpc regionServerReport(RegionServerReportRequest) + rpc RegionServerReport(RegionServerReportRequest) returns(RegionServerReportResponse); /** * Called by a region server to report a fatal error that is causing it to * abort. */ - rpc reportRSFatalError(ReportRSFatalErrorRequest) + rpc ReportRSFatalError(ReportRSFatalErrorRequest) returns(ReportRSFatalErrorResponse); /** Called to get the sequence id of the last MemStore entry flushed to an * HFile for a specified region. Used by the region server to speed up * log splitting. */ - rpc getLastFlushedSequenceId(GetLastFlushedSequenceIdRequest) + rpc GetLastFlushedSequenceId(GetLastFlushedSequenceIdRequest) returns(GetLastFlushedSequenceIdResponse); } diff --git hbase-protocol/src/main/protobuf/RowProcessor.proto hbase-protocol/src/main/protobuf/RowProcessor.proto index ff2588a..7f6054c 100644 --- hbase-protocol/src/main/protobuf/RowProcessor.proto +++ hbase-protocol/src/main/protobuf/RowProcessor.proto @@ -27,15 +27,15 @@ option java_generate_equals_and_hash = true; option optimize_for = SPEED; message RowProcessorRequest { - required string rowProcessorClassName = 1; - optional string rowProcessorInitializerMessageName = 2; - optional bytes rowProcessorInitializerMessage = 3; + required string row_processor_class_name = 1; + optional string row_processor_initializer_message_name = 2; + optional bytes row_processor_initializer_message = 3; } message RowProcessorResult { - required bytes rowProcessorResult = 1; + required bytes row_processor_result = 1; } service RowProcessorService { - rpc process (RowProcessorRequest) returns (RowProcessorResult); -} \ No newline at end of file + rpc Process (RowProcessorRequest) returns (RowProcessorResult); +} diff --git hbase-protocol/src/main/protobuf/SecureBulkLoad.proto hbase-protocol/src/main/protobuf/SecureBulkLoad.proto index 40fe789..7d5604b 100644 --- hbase-protocol/src/main/protobuf/SecureBulkLoad.proto +++ hbase-protocol/src/main/protobuf/SecureBulkLoad.proto @@ -26,10 +26,10 @@ import 'hbase.proto'; import 'Client.proto'; message SecureBulkLoadHFilesRequest { - repeated BulkLoadHFileRequest.FamilyPath familyPath = 1; - optional bool assignSeqNum = 2; - required DelegationTokenProto fsToken = 3; - required string bulkToken = 4; + repeated BulkLoadHFileRequest.FamilyPath family_path = 1; + optional bool assign_seq_num = 2; + required DelegationTokenProto fs_token = 3; + required string bulk_token = 4; } message SecureBulkLoadHFilesResponse { @@ -44,15 +44,15 @@ message DelegationTokenProto { } message PrepareBulkLoadRequest { - required bytes tableName = 1; + required bytes table_name = 1; } message PrepareBulkLoadResponse { - required string bulkToken = 1; + required string bulk_token = 1; } message CleanupBulkLoadRequest { - required string bulkToken = 1; + required string bulk_token = 1; } @@ -60,12 +60,12 @@ message CleanupBulkLoadResponse { } service SecureBulkLoadService { - rpc prepareBulkLoad(PrepareBulkLoadRequest) + rpc PrepareBulkLoad(PrepareBulkLoadRequest) returns (PrepareBulkLoadResponse); - rpc secureBulkLoadHFiles(SecureBulkLoadHFilesRequest) + rpc SecureBulkLoadHFiles(SecureBulkLoadHFilesRequest) returns (SecureBulkLoadHFilesResponse); - rpc cleanupBulkLoad(CleanupBulkLoadRequest) + rpc CleanupBulkLoad(CleanupBulkLoadRequest) returns (CleanupBulkLoadResponse); } diff --git hbase-protocol/src/main/protobuf/Tracing.proto hbase-protocol/src/main/protobuf/Tracing.proto index 8e5babd..688f7c5 100644 --- hbase-protocol/src/main/protobuf/Tracing.proto +++ hbase-protocol/src/main/protobuf/Tracing.proto @@ -26,6 +26,6 @@ option optimize_for = SPEED; //the id of the current span when this message was sent, so we know //what span caused the new span we will create when this message is received. message RPCTInfo { - optional int64 traceId = 1; - optional int64 parentId = 2; + optional int64 trace_id = 1; + optional int64 parent_id = 2; } diff --git hbase-protocol/src/main/protobuf/WAL.proto hbase-protocol/src/main/protobuf/WAL.proto index 5635c60..b6f6da6 100644 --- hbase-protocol/src/main/protobuf/WAL.proto +++ hbase-protocol/src/main/protobuf/WAL.proto @@ -24,21 +24,21 @@ option optimize_for = SPEED; import "hbase.proto"; message WALHeader { - optional bool hasCompression = 1; + optional bool has_compression = 1; } // Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs message WALKey { - required bytes encodedRegionName = 1; - required bytes tableName = 2; - required uint64 logSequenceNumber = 3; - required uint64 writeTime = 4; - optional UUID clusterId = 5; + required bytes encoded_region_name = 1; + required bytes table_name = 2; + required uint64 log_sequence_number = 3; + required uint64 write_time = 4; + optional UUID cluster_id = 5; repeated FamilyScope scopes = 6; - optional uint32 followingKvCount = 7; + optional uint32 following_kv_count = 7; /* - optional CustomEntryType customEntryType = 8; + optional CustomEntryType custom_entry_type = 8; enum CustomEntryType { COMPACTION = 0; @@ -53,7 +53,7 @@ enum ScopeType { message FamilyScope { required bytes family = 1; - required ScopeType scopeType = 2; + required ScopeType scope_type = 2; } /** @@ -67,12 +67,12 @@ message FamilyScope { * the * compaction should we fail the WAL write. */ message CompactionDescriptor { - required bytes tableName = 1; // TODO: WALKey already stores these, might remove - required bytes encodedRegionName = 2; - required bytes familyName = 3; - repeated string compactionInput = 4; - repeated string compactionOutput = 5; - required string storeHomeDir = 6; + required bytes table_name = 1; // TODO: WALKey already stores these, might remove + required bytes encoded_region_name = 2; + required bytes family_name = 3; + repeated string compaction_input = 4; + repeated string compaction_output = 5; + required string store_home_dir = 6; } /** diff --git hbase-protocol/src/main/protobuf/ZooKeeper.proto hbase-protocol/src/main/protobuf/ZooKeeper.proto index 3536694..ef3dcbf 100644 --- hbase-protocol/src/main/protobuf/ZooKeeper.proto +++ hbase-protocol/src/main/protobuf/ZooKeeper.proto @@ -36,7 +36,7 @@ message MetaRegionServer { // The major version of the rpc the server speaks. This is used so that // clients connecting to the cluster can have prior knowledge of what version // to send to a RegionServer. AsyncHBase will use this to detect versions. - optional uint32 rpcVersion = 2; + optional uint32 rpc_version = 2; } /** @@ -46,7 +46,7 @@ message Master { // The ServerName of the current Master required ServerName master = 1; // Major RPC version so that clients can know what version the master can accept. - optional uint32 rpcVersion = 2; + optional uint32 rpc_version = 2; } /** @@ -54,8 +54,8 @@ message Master { */ message ClusterUp { // If this znode is present, cluster is up. Currently - // the data is cluster startDate. - required string startDate = 1; + // the data is cluster start_date. + required string start_date = 1; } /** @@ -64,12 +64,12 @@ message ClusterUp { */ message RegionTransition { // Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() - required uint32 eventTypeCode = 1; + required uint32 event_type_code = 1; // Full regionname in bytes - required bytes regionName = 2; - required uint64 createTime = 3; + required bytes region_name = 2; + required uint64 create_time = 3; // The region server where the transition will happen or is happening - required ServerName serverName = 4; + required ServerName server_name = 4; optional bytes payload = 5; } @@ -86,7 +86,7 @@ message SplitLogTask { ERR = 4; } required State state = 1; - required ServerName serverName = 2; + required ServerName server_name = 2; } /** @@ -110,7 +110,7 @@ message Table { * Used by replication. Holds a replication peer key. */ message ReplicationPeer { - // clusterKey is the concatenation of the slave cluster's + // clusterkey is the concatenation of the slave cluster's // hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent required string clusterkey = 1; } @@ -137,27 +137,27 @@ message ReplicationHLogPosition { * Used by replication. Used to lock a region server during failover. */ message ReplicationLock { - required string lockOwner = 1; + required string lock_owner = 1; } /** * Metadata associated with a table lock in zookeeper */ message TableLock { - optional bytes tableName = 1; - optional ServerName lockOwner = 2; - optional int64 threadId = 3; - optional bool isShared = 4; + optional bytes table_name = 1; + optional ServerName lock_owner = 2; + optional int64 thread_id = 3; + optional bool is_shared = 4; optional string purpose = 5; - optional int64 createTime = 6; + optional int64 create_time = 6; } /** * sequence Id of a store */ message StoreSequenceId { - required bytes familyName = 1; - required uint64 sequenceId = 2; + required bytes family_name = 1; + required uint64 sequence_id = 2; } /** @@ -165,6 +165,6 @@ message StoreSequenceId { * list sequence ids of the region's stores */ message RegionStoreSequenceIds { - required uint64 lastFlushedSequenceId = 1; - repeated StoreSequenceId storeSequenceId = 2; + required uint64 last_flushed_sequence_id = 1; + repeated StoreSequenceId store_sequence_id = 2; } diff --git hbase-protocol/src/main/protobuf/hbase.proto hbase-protocol/src/main/protobuf/hbase.proto index 8846b89..9dfe084 100644 --- hbase-protocol/src/main/protobuf/hbase.proto +++ hbase-protocol/src/main/protobuf/hbase.proto @@ -32,7 +32,7 @@ import "Cell.proto"; message TableSchema { optional bytes name = 1; repeated BytesBytesPair attributes = 2; - repeated ColumnFamilySchema columnFamilies = 3; + repeated ColumnFamilySchema column_families = 3; repeated NameStringPair configuration = 4; } @@ -50,10 +50,10 @@ message ColumnFamilySchema { * Protocol buffer version of HRegionInfo. */ message RegionInfo { - required uint64 regionId = 1; - required bytes tableName = 2; - optional bytes startKey = 3; - optional bytes endKey = 4; + required uint64 region_id = 1; + required bytes table_name = 2; + optional bytes start_key = 3; + optional bytes end_key = 4; optional bool offline = 5; optional bool split = 6; optional bool recovering = 7; @@ -63,7 +63,7 @@ message RegionInfo { * Protocol buffer for favored nodes */ message FavoredNodes { - repeated ServerName favoredNode = 1; + repeated ServerName favored_node = 1; } /** @@ -111,9 +111,9 @@ enum CompareType { * Protocol buffer version of ServerName */ message ServerName { - required string hostName = 1; + required string host_name = 1; optional uint32 port = 2; - optional uint64 startCode = 3; + optional uint64 start_code = 3; } // Comment data structures @@ -148,7 +148,7 @@ message NameInt64Pair { message SnapshotDescription { required string name = 1; optional string table = 2; // not needed for delete, but checked for in taking snapshot - optional int64 creationTime = 3 [default = 0]; + optional int64 creation_time = 3 [default = 0]; enum Type { DISABLED = 0; FLUSH = 1; @@ -161,15 +161,15 @@ message EmptyMsg { } message LongMsg { - required int64 longMsg = 1; + required int64 long_msg = 1; } message BigDecimalMsg { - required bytes bigdecimalMsg = 1; + required bytes bigdecimal_msg = 1; } message UUID { - required uint64 leastSigBits = 1; - required uint64 mostSigBits = 2; + required uint64 least_sig_bits = 1; + required uint64 most_sig_bits = 2; } diff --git hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java index 30bb8af..f48bc5a 100644 --- hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java +++ hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java @@ -133,7 +133,7 @@ public class TokenProvider implements AuthenticationProtos.AuthenticationService } @Override - public void whoami(RpcController controller, AuthenticationProtos.WhoAmIRequest request, + public void whoAmI(RpcController controller, AuthenticationProtos.WhoAmIRequest request, RpcCallback done) { User requestUser = RequestContext.getRequestUser(); AuthenticationProtos.WhoAmIResponse.Builder response = diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java index 7b676ee..c136479 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java @@ -274,15 +274,15 @@ public class TestTokenAuthentication { } @Override - public AuthenticationProtos.WhoAmIResponse whoami( + public AuthenticationProtos.WhoAmIResponse whoAmI( RpcController controller, AuthenticationProtos.WhoAmIRequest request) throws ServiceException { - LOG.debug("whoami() request from "+RequestContext.getRequestUserName()); + LOG.debug("whoAmI() request from "+RequestContext.getRequestUserName()); // ignore passed in controller -- it's always null ServerRpcController serverController = new ServerRpcController(); BlockingRpcCallback callback = new BlockingRpcCallback(); - whoami(serverController, request, callback); + whoAmI(serverController, request, callback); try { serverController.checkFailed(); return callback.get(); @@ -383,7 +383,7 @@ public class TestTokenAuthentication { AuthenticationProtos.AuthenticationService.BlockingInterface stub = AuthenticationProtos.AuthenticationService.newBlockingStub(channel); AuthenticationProtos.WhoAmIResponse response = - stub.whoami(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance()); + stub.whoAmI(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance()); String myname = response.getUsername(); assertEquals("testuser", myname); String authMethod = response.getAuthMethod();