diff --git a/hbase-assembly/pom.xml b/hbase-assembly/pom.xml index a5f869f..3a3c5c7 100644 --- a/hbase-assembly/pom.xml +++ b/hbase-assembly/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. hbase-assembly diff --git a/hbase-client/pom.xml b/hbase-client/pom.xml index b0f9928..75401ee 100644 --- a/hbase-client/pom.xml +++ b/hbase-client/pom.xml @@ -24,7 +24,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java index 53704ea..236304c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java @@ -26,7 +26,7 @@ import java.util.Collections; import java.util.HashMap; import java.util.Map; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.master.RegionState; @@ -334,7 +334,7 @@ public class ClusterStatus extends VersionedWritable { ClusterStatusProtos.RegionState rs = rit.getValue().convert(); RegionSpecifier.Builder spec = RegionSpecifier.newBuilder().setType(RegionSpecifierType.REGION_NAME); - spec.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(rit.getKey()))); + spec.setValue(ByteStringer.wrap(Bytes.toBytes(rit.getKey()))); RegionInTransition pbRIT = RegionInTransition.newBuilder().setSpec(spec.build()).setRegionState(rs).build(); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java index 2a5a70c..e7cdddb 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -45,7 +45,7 @@ import org.apache.hadoop.io.Text; import org.apache.hadoop.io.WritableComparable; import com.google.common.base.Preconditions; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; /** @@ -1272,11 +1272,11 @@ public class HColumnDescriptor implements WritableComparable */ public ColumnFamilySchema convert() { ColumnFamilySchema.Builder builder = ColumnFamilySchema.newBuilder(); - builder.setName(HBaseZeroCopyByteString.wrap(getName())); + builder.setName(ByteStringer.wrap(getName())); for (Map.Entry e: this.values.entrySet()) { BytesBytesPair.Builder aBuilder = BytesBytesPair.newBuilder(); - aBuilder.setFirst(HBaseZeroCopyByteString.wrap(e.getKey().get())); - aBuilder.setSecond(HBaseZeroCopyByteString.wrap(e.getValue().get())); + aBuilder.setFirst(ByteStringer.wrap(e.getKey().get())); + aBuilder.setSecond(ByteStringer.wrap(e.getValue().get())); builder.addAttributes(aBuilder.build()); } for (Map.Entry e : this.configuration.entrySet()) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java index afd1d04..502688c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java @@ -29,7 +29,7 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.List; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -842,10 +842,10 @@ public class HRegionInfo implements Comparable { builder.setTableName(ProtobufUtil.toProtoTableName(info.getTable())); builder.setRegionId(info.getRegionId()); if (info.getStartKey() != null) { - builder.setStartKey(HBaseZeroCopyByteString.wrap(info.getStartKey())); + builder.setStartKey(ByteStringer.wrap(info.getStartKey())); } if (info.getEndKey() != null) { - builder.setEndKey(HBaseZeroCopyByteString.wrap(info.getEndKey())); + builder.setEndKey(ByteStringer.wrap(info.getEndKey())); } builder.setOffline(info.isOffline()); builder.setSplit(info.isSplit()); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java index d8dad03..702618c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java @@ -34,7 +34,7 @@ import java.util.TreeMap; import java.util.TreeSet; import java.util.regex.Matcher; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -1432,8 +1432,8 @@ public class HTableDescriptor implements WritableComparable { builder.setTableName(ProtobufUtil.toProtoTableName(getTableName())); for (Map.Entry e: this.values.entrySet()) { BytesBytesPair.Builder aBuilder = BytesBytesPair.newBuilder(); - aBuilder.setFirst(HBaseZeroCopyByteString.wrap(e.getKey().get())); - aBuilder.setSecond(HBaseZeroCopyByteString.wrap(e.getValue().get())); + aBuilder.setFirst(ByteStringer.wrap(e.getKey().get())); + aBuilder.setSecond(ByteStringer.wrap(e.getValue().get())); builder.addAttributes(aBuilder.build()); } for (HColumnDescriptor hcd: getColumnFamilies()) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionTransition.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionTransition.java index 0632825..7e7f5c1 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionTransition.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/RegionTransition.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.classification.InterfaceAudience; @@ -104,10 +104,10 @@ public class RegionTransition { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(). setHostName(sn.getHostname()).setPort(sn.getPort()).setStartCode(sn.getStartcode()).build(); ZooKeeperProtos.RegionTransition.Builder builder = ZooKeeperProtos.RegionTransition.newBuilder(). - setEventTypeCode(type.getCode()).setRegionName(HBaseZeroCopyByteString.wrap(regionName)). + setEventTypeCode(type.getCode()).setRegionName(ByteStringer.wrap(regionName)). setServerName(pbsn); builder.setCreateTime(System.currentTimeMillis()); - if (payload != null) builder.setPayload(HBaseZeroCopyByteString.wrap(payload)); + if (payload != null) builder.setPayload(ByteStringer.wrap(payload)); return new RegionTransition(builder.build()); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.java index cc1fcfd..f52b739 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.java @@ -22,7 +22,7 @@ import java.io.IOException; import java.math.BigDecimal; import java.math.RoundingMode; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; @@ -121,7 +121,7 @@ public class BigDecimalColumnInterpreter extends ColumnInterpreter(); for(Pair el: familyPaths) { protoFamilyPaths.add(ClientProtos.BulkLoadHFileRequest.FamilyPath.newBuilder() - .setFamily(HBaseZeroCopyByteString.wrap(el.getFirst())) + .setFamily(ByteStringer.wrap(el.getFirst())) .setPath(el.getSecond()).build()); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java index 73d735a..fc4e234 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ByteArrayComparable.java @@ -18,7 +18,7 @@ */ package org.apache.hadoop.hbase.filter; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.exceptions.DeserializationException; @@ -53,7 +53,7 @@ public abstract class ByteArrayComparable implements Comparable { ComparatorProtos.ByteArrayComparable convert() { ComparatorProtos.ByteArrayComparable.Builder builder = ComparatorProtos.ByteArrayComparable.newBuilder(); - if (value != null) builder.setValue(HBaseZeroCopyByteString.wrap(value)); + if (value != null) builder.setValue(ByteStringer.wrap(value)); return builder.build(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java index 8094395..487d512 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.filter; import java.util.ArrayList; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Cell; @@ -175,7 +175,7 @@ public class ColumnPaginationFilter extends FilterBase builder.setOffset(this.offset); } if (this.columnOffset != null) { - builder.setColumnOffset(HBaseZeroCopyByteString.wrap(this.columnOffset)); + builder.setColumnOffset(ByteStringer.wrap(this.columnOffset)); } return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java index a6873d9..012a96d 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hbase.filter; import java.util.ArrayList; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Cell; @@ -95,7 +95,7 @@ public class ColumnPrefixFilter extends FilterBase { public byte [] toByteArray() { FilterProtos.ColumnPrefixFilter.Builder builder = FilterProtos.ColumnPrefixFilter.newBuilder(); - if (this.prefix != null) builder.setPrefix(HBaseZeroCopyByteString.wrap(this.prefix)); + if (this.prefix != null) builder.setPrefix(ByteStringer.wrap(this.prefix)); return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java index 85c96f0..408d665 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java @@ -22,7 +22,7 @@ package org.apache.hadoop.hbase.filter; import static org.apache.hadoop.hbase.util.Bytes.len; import com.google.common.base.Preconditions; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.classification.InterfaceAudience; @@ -173,9 +173,9 @@ public class ColumnRangeFilter extends FilterBase { public byte [] toByteArray() { FilterProtos.ColumnRangeFilter.Builder builder = FilterProtos.ColumnRangeFilter.newBuilder(); - if (this.minColumn != null) builder.setMinColumn(HBaseZeroCopyByteString.wrap(this.minColumn)); + if (this.minColumn != null) builder.setMinColumn(ByteStringer.wrap(this.minColumn)); builder.setMinColumnInclusive(this.minColumnInclusive); - if (this.maxColumn != null) builder.setMaxColumn(HBaseZeroCopyByteString.wrap(this.maxColumn)); + if (this.maxColumn != null) builder.setMaxColumn(ByteStringer.wrap(this.maxColumn)); builder.setMaxColumnInclusive(this.maxColumnInclusive); return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java index 32a3dcd..aa0586d 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java @@ -25,7 +25,7 @@ import java.util.Iterator; import java.util.List; import java.util.Set; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Cell; @@ -225,10 +225,10 @@ public class DependentColumnFilter extends CompareFilter { FilterProtos.DependentColumnFilter.newBuilder(); builder.setCompareFilter(super.convert()); if (this.columnFamily != null) { - builder.setColumnFamily(HBaseZeroCopyByteString.wrap(this.columnFamily)); + builder.setColumnFamily(ByteStringer.wrap(this.columnFamily)); } if (this.columnQualifier != null) { - builder.setColumnQualifier(HBaseZeroCopyByteString.wrap(this.columnQualifier)); + builder.setColumnQualifier(ByteStringer.wrap(this.columnQualifier)); } builder.setDropDependentColumn(this.dropDependentColumn); return builder.build().toByteArray(); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java index edee29a..9bad808 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyValueMatchingQualifiersFilter.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.filter; import com.google.protobuf.ByteString; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.classification.InterfaceAudience; @@ -89,7 +89,7 @@ public class FirstKeyValueMatchingQualifiersFilter extends FirstKeyOnlyFilter { FilterProtos.FirstKeyValueMatchingQualifiersFilter.Builder builder = FilterProtos.FirstKeyValueMatchingQualifiersFilter.newBuilder(); for (byte[] qualifier : qualifiers) { - if (qualifier != null) builder.addQualifiers(HBaseZeroCopyByteString.wrap(qualifier)); + if (qualifier != null) builder.addQualifiers(ByteStringer.wrap(qualifier)); } return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java index ee3ece2..d565f31 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java @@ -17,7 +17,7 @@ */ package org.apache.hadoop.hbase.filter; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.classification.InterfaceAudience; @@ -147,8 +147,8 @@ public class FuzzyRowFilter extends FilterBase { FilterProtos.FuzzyRowFilter.newBuilder(); for (Pair fuzzyData : fuzzyKeysData) { BytesBytesPair.Builder bbpBuilder = BytesBytesPair.newBuilder(); - bbpBuilder.setFirst(HBaseZeroCopyByteString.wrap(fuzzyData.getFirst())); - bbpBuilder.setSecond(HBaseZeroCopyByteString.wrap(fuzzyData.getSecond())); + bbpBuilder.setFirst(ByteStringer.wrap(fuzzyData.getFirst())); + bbpBuilder.setSecond(ByteStringer.wrap(fuzzyData.getSecond())); builder.addFuzzyKeysData(bbpBuilder); } return builder.build().toByteArray(); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java index 2c03067..bceb1a3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hbase.filter; import java.util.ArrayList; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Cell; @@ -94,7 +94,7 @@ public class InclusiveStopFilter extends FilterBase { public byte [] toByteArray() { FilterProtos.InclusiveStopFilter.Builder builder = FilterProtos.InclusiveStopFilter.newBuilder(); - if (this.stopRowKey != null) builder.setStopRowKey(HBaseZeroCopyByteString.wrap(this.stopRowKey)); + if (this.stopRowKey != null) builder.setStopRowKey(ByteStringer.wrap(this.stopRowKey)); return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java index 88ea67b..e80b094 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hbase.filter; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.classification.InterfaceAudience; @@ -114,7 +114,7 @@ public class MultipleColumnPrefixFilter extends FilterBase { FilterProtos.MultipleColumnPrefixFilter.Builder builder = FilterProtos.MultipleColumnPrefixFilter.newBuilder(); for (byte [] element : sortedPrefixes) { - if (element != null) builder.addSortedPrefixes(HBaseZeroCopyByteString.wrap(element)); + if (element != null) builder.addSortedPrefixes(ByteStringer.wrap(element)); } return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java index 29bdcde..c0c8396 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.filter; import com.google.common.base.Preconditions; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.InvalidProtocolBufferException; import org.apache.hadoop.classification.InterfaceAudience; @@ -98,7 +98,7 @@ public class PrefixFilter extends FilterBase { public byte [] toByteArray() { FilterProtos.PrefixFilter.Builder builder = FilterProtos.PrefixFilter.newBuilder(); - if (this.prefix != null) builder.setPrefix(HBaseZeroCopyByteString.wrap(this.prefix)); + if (this.prefix != null) builder.setPrefix(ByteStringer.wrap(this.prefix)); return builder.build().toByteArray(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java index 771efeb..21a2207 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java @@ -22,7 +22,7 @@ package org.apache.hadoop.hbase.filter; import java.io.IOException; import java.util.ArrayList; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -307,10 +307,10 @@ public class SingleColumnValueFilter extends FilterBase { FilterProtos.SingleColumnValueFilter.Builder builder = FilterProtos.SingleColumnValueFilter.newBuilder(); if (this.columnFamily != null) { - builder.setColumnFamily(HBaseZeroCopyByteString.wrap(this.columnFamily)); + builder.setColumnFamily(ByteStringer.wrap(this.columnFamily)); } if (this.columnQualifier != null) { - builder.setColumnQualifier(HBaseZeroCopyByteString.wrap(this.columnQualifier)); + builder.setColumnQualifier(ByteStringer.wrap(this.columnQualifier)); } HBaseProtos.CompareType compareOp = CompareType.valueOf(this.compareOp.name()); builder.setCompareOp(compareOp); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java index 97fdb6f..c60b6b8 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/MasterCoprocessorRpcChannel.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -61,7 +61,7 @@ public class MasterCoprocessorRpcChannel extends CoprocessorRpcChannel{ final ClientProtos.CoprocessorServiceCall call = ClientProtos.CoprocessorServiceCall.newBuilder() - .setRow(HBaseZeroCopyByteString.wrap(HConstants.EMPTY_BYTE_ARRAY)) + .setRow(ByteStringer.wrap(HConstants.EMPTY_BYTE_ARRAY)) .setServiceName(method.getService().getFullName()) .setMethodName(method.getName()) .setRequest(request.toByteString()).build(); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java index de1259d..818268b 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RegionCoprocessorRpcChannel.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -80,7 +80,7 @@ public class RegionCoprocessorRpcChannel extends CoprocessorRpcChannel{ final ClientProtos.CoprocessorServiceCall call = ClientProtos.CoprocessorServiceCall.newBuilder() - .setRow(HBaseZeroCopyByteString.wrap(row)) + .setRow(ByteStringer.wrap(row)) .setServiceName(method.getService().getFullName()) .setMethodName(method.getName()) .setRequest(request.toByteString()).build(); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index c28ebfd..d381ec7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -36,7 +36,6 @@ import java.util.Map; import java.util.Map.Entry; import java.util.NavigableSet; -import com.google.protobuf.HBaseZeroCopyByteString; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.Cell; @@ -124,6 +123,7 @@ import org.apache.hadoop.hbase.security.access.UserPermission; import org.apache.hadoop.hbase.security.token.AuthenticationTokenIdentifier; import org.apache.hadoop.hbase.security.visibility.Authorizations; import org.apache.hadoop.hbase.security.visibility.CellVisibility; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.DynamicClassLoader; import org.apache.hadoop.hbase.util.ExceptionUtil; @@ -825,17 +825,17 @@ public final class ProtobufUtil { NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder(); for (Map.Entry attribute: attributes.entrySet()) { attributeBuilder.setName(attribute.getKey()); - attributeBuilder.setValue(HBaseZeroCopyByteString.wrap(attribute.getValue())); + attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue())); scanBuilder.addAttribute(attributeBuilder.build()); } } byte[] startRow = scan.getStartRow(); if (startRow != null && startRow.length > 0) { - scanBuilder.setStartRow(HBaseZeroCopyByteString.wrap(startRow)); + scanBuilder.setStartRow(ByteStringer.wrap(startRow)); } byte[] stopRow = scan.getStopRow(); if (stopRow != null && stopRow.length > 0) { - scanBuilder.setStopRow(HBaseZeroCopyByteString.wrap(stopRow)); + scanBuilder.setStopRow(ByteStringer.wrap(stopRow)); } if (scan.hasFilter()) { scanBuilder.setFilter(ProtobufUtil.toFilter(scan.getFilter())); @@ -844,12 +844,12 @@ public final class ProtobufUtil { Column.Builder columnBuilder = Column.newBuilder(); for (Map.Entry> family: scan.getFamilyMap().entrySet()) { - columnBuilder.setFamily(HBaseZeroCopyByteString.wrap(family.getKey())); + columnBuilder.setFamily(ByteStringer.wrap(family.getKey())); NavigableSet qualifiers = family.getValue(); columnBuilder.clearQualifier(); if (qualifiers != null && qualifiers.size() > 0) { for (byte [] qualifier: qualifiers) { - columnBuilder.addQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + columnBuilder.addQualifier(ByteStringer.wrap(qualifier)); } } scanBuilder.addColumn(columnBuilder.build()); @@ -957,7 +957,7 @@ public final class ProtobufUtil { final Get get) throws IOException { ClientProtos.Get.Builder builder = ClientProtos.Get.newBuilder(); - builder.setRow(HBaseZeroCopyByteString.wrap(get.getRow())); + builder.setRow(ByteStringer.wrap(get.getRow())); builder.setCacheBlocks(get.getCacheBlocks()); builder.setMaxVersions(get.getMaxVersions()); if (get.getFilter() != null) { @@ -976,7 +976,7 @@ public final class ProtobufUtil { NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder(); for (Map.Entry attribute: attributes.entrySet()) { attributeBuilder.setName(attribute.getKey()); - attributeBuilder.setValue(HBaseZeroCopyByteString.wrap(attribute.getValue())); + attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue())); builder.addAttribute(attributeBuilder.build()); } } @@ -985,11 +985,11 @@ public final class ProtobufUtil { Map> families = get.getFamilyMap(); for (Map.Entry> family: families.entrySet()) { NavigableSet qualifiers = family.getValue(); - columnBuilder.setFamily(HBaseZeroCopyByteString.wrap(family.getKey())); + columnBuilder.setFamily(ByteStringer.wrap(family.getKey())); columnBuilder.clearQualifier(); if (qualifiers != null && qualifiers.size() > 0) { for (byte[] qualifier: qualifiers) { - columnBuilder.addQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + columnBuilder.addQualifier(ByteStringer.wrap(qualifier)); } } builder.addColumn(columnBuilder.build()); @@ -1018,7 +1018,7 @@ public final class ProtobufUtil { */ public static MutationProto toMutation( final Increment increment, final MutationProto.Builder builder, long nonce) { - builder.setRow(HBaseZeroCopyByteString.wrap(increment.getRow())); + builder.setRow(ByteStringer.wrap(increment.getRow())); builder.setMutateType(MutationType.INCREMENT); builder.setDurability(toDurability(increment.getDurability())); if (nonce != HConstants.NO_NONCE) { @@ -1035,18 +1035,18 @@ public final class ProtobufUtil { ColumnValue.Builder columnBuilder = ColumnValue.newBuilder(); QualifierValue.Builder valueBuilder = QualifierValue.newBuilder(); for (Map.Entry> family: increment.getFamilyCellMap().entrySet()) { - columnBuilder.setFamily(HBaseZeroCopyByteString.wrap(family.getKey())); + columnBuilder.setFamily(ByteStringer.wrap(family.getKey())); columnBuilder.clearQualifierValue(); List values = family.getValue(); if (values != null && values.size() > 0) { for (Cell cell: values) { KeyValue kv = KeyValueUtil.ensureKeyValue(cell); - valueBuilder.setQualifier(HBaseZeroCopyByteString.wrap( + valueBuilder.setQualifier(ByteStringer.wrap( kv.getQualifierArray(), kv.getQualifierOffset(), kv.getQualifierLength())); - valueBuilder.setValue(HBaseZeroCopyByteString.wrap( + valueBuilder.setValue(ByteStringer.wrap( kv.getValueArray(), kv.getValueOffset(), kv.getValueLength())); if (kv.getTagsLength() > 0) { - valueBuilder.setTags(HBaseZeroCopyByteString.wrap(kv.getTagsArray(), + valueBuilder.setTags(ByteStringer.wrap(kv.getTagsArray(), kv.getTagsOffset(), kv.getTagsLength())); } columnBuilder.addQualifierValue(valueBuilder.build()); @@ -1059,7 +1059,7 @@ public final class ProtobufUtil { NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder(); for (Map.Entry attribute : attributes.entrySet()) { attributeBuilder.setName(attribute.getKey()); - attributeBuilder.setValue(HBaseZeroCopyByteString.wrap(attribute.getValue())); + attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue())); builder.addAttribute(attributeBuilder.build()); } } @@ -1100,16 +1100,16 @@ public final class ProtobufUtil { QualifierValue.Builder valueBuilder = QualifierValue.newBuilder(); for (Map.Entry> family: mutation.getFamilyCellMap().entrySet()) { columnBuilder.clear(); - columnBuilder.setFamily(HBaseZeroCopyByteString.wrap(family.getKey())); + columnBuilder.setFamily(ByteStringer.wrap(family.getKey())); for (Cell cell: family.getValue()) { KeyValue kv = KeyValueUtil.ensureKeyValue(cell); - valueBuilder.setQualifier(HBaseZeroCopyByteString.wrap( + valueBuilder.setQualifier(ByteStringer.wrap( kv.getQualifierArray(), kv.getQualifierOffset(), kv.getQualifierLength())); - valueBuilder.setValue(HBaseZeroCopyByteString.wrap( + valueBuilder.setValue(ByteStringer.wrap( kv.getValueArray(), kv.getValueOffset(), kv.getValueLength())); valueBuilder.setTimestamp(kv.getTimestamp()); if(cell.getTagsLength() > 0) { - valueBuilder.setTags(HBaseZeroCopyByteString.wrap(kv.getTagsArray(), kv.getTagsOffset(), + valueBuilder.setTags(ByteStringer.wrap(kv.getTagsArray(), kv.getTagsOffset(), kv.getTagsLength())); } if (type == MutationType.DELETE) { @@ -1170,7 +1170,7 @@ public final class ProtobufUtil { */ private static MutationProto.Builder getMutationBuilderAndSetCommonFields(final MutationType type, final Mutation mutation, MutationProto.Builder builder) { - builder.setRow(HBaseZeroCopyByteString.wrap(mutation.getRow())); + builder.setRow(ByteStringer.wrap(mutation.getRow())); builder.setMutateType(type); builder.setDurability(toDurability(mutation.getDurability())); builder.setTimestamp(mutation.getTimeStamp()); @@ -1179,7 +1179,7 @@ public final class ProtobufUtil { NameBytesPair.Builder attributeBuilder = NameBytesPair.newBuilder(); for (Map.Entry attribute: attributes.entrySet()) { attributeBuilder.setName(attribute.getKey()); - attributeBuilder.setValue(HBaseZeroCopyByteString.wrap(attribute.getValue())); + attributeBuilder.setValue(ByteStringer.wrap(attribute.getValue())); builder.addAttribute(attributeBuilder.build()); } } @@ -1310,7 +1310,7 @@ public final class ProtobufUtil { public static ComparatorProtos.Comparator toComparator(ByteArrayComparable comparator) { ComparatorProtos.Comparator.Builder builder = ComparatorProtos.Comparator.newBuilder(); builder.setName(comparator.getClass().getName()); - builder.setSerializedComparator(HBaseZeroCopyByteString.wrap(comparator.toByteArray())); + builder.setSerializedComparator(ByteStringer.wrap(comparator.toByteArray())); return builder.build(); } @@ -1372,7 +1372,7 @@ public final class ProtobufUtil { public static FilterProtos.Filter toFilter(Filter filter) throws IOException { FilterProtos.Filter.Builder builder = FilterProtos.Filter.newBuilder(); builder.setName(filter.getClass().getName()); - builder.setSerializedFilter(HBaseZeroCopyByteString.wrap(filter.toByteArray())); + builder.setSerializedFilter(ByteStringer.wrap(filter.toByteArray())); return builder.build(); } @@ -1915,10 +1915,10 @@ public final class ProtobufUtil { AccessControlProtos.TablePermission.newBuilder(); builder.setTableName(ProtobufUtil.toProtoTableName(tablePerm.getTableName())); if (tablePerm.hasFamily()) { - builder.setFamily(HBaseZeroCopyByteString.wrap(tablePerm.getFamily())); + builder.setFamily(ByteStringer.wrap(tablePerm.getFamily())); } if (tablePerm.hasQualifier()) { - builder.setQualifier(HBaseZeroCopyByteString.wrap(tablePerm.getQualifier())); + builder.setQualifier(ByteStringer.wrap(tablePerm.getQualifier())); } Permission.Action actions[] = perm.getActions(); if (actions != null) { @@ -2014,7 +2014,7 @@ public final class ProtobufUtil { */ public static AccessControlProtos.UserPermission toUserPermission(UserPermission perm) { return AccessControlProtos.UserPermission.newBuilder() - .setUser(HBaseZeroCopyByteString.wrap(perm.getUser())) + .setUser(ByteStringer.wrap(perm.getUser())) .setPermission(toPermission(perm)) .build(); } @@ -2270,7 +2270,7 @@ public final class ProtobufUtil { AccessControlProtos.GetUserPermissionsRequest.Builder builder = AccessControlProtos.GetUserPermissionsRequest.newBuilder(); if (namespace != null) { - builder.setNamespaceName(HBaseZeroCopyByteString.wrap(namespace)); + builder.setNamespaceName(ByteStringer.wrap(namespace)); } builder.setType(AccessControlProtos.Permission.Type.Namespace); AccessControlProtos.GetUserPermissionsRequest request = builder.build(); @@ -2314,8 +2314,8 @@ public final class ProtobufUtil { */ public static AuthenticationProtos.Token toToken(Token token) { AuthenticationProtos.Token.Builder builder = AuthenticationProtos.Token.newBuilder(); - builder.setIdentifier(HBaseZeroCopyByteString.wrap(token.getIdentifier())); - builder.setPassword(HBaseZeroCopyByteString.wrap(token.getPassword())); + builder.setIdentifier(ByteStringer.wrap(token.getIdentifier())); + builder.setPassword(ByteStringer.wrap(token.getPassword())); if (token.getService() != null) { builder.setService(ByteString.copyFromUtf8(token.getService().toString())); } @@ -2412,15 +2412,15 @@ public final class ProtobufUtil { // Doing this is going to kill us if we do it for all data passed. // St.Ack 20121205 CellProtos.Cell.Builder kvbuilder = CellProtos.Cell.newBuilder(); - kvbuilder.setRow(HBaseZeroCopyByteString.wrap(kv.getRowArray(), kv.getRowOffset(), + kvbuilder.setRow(ByteStringer.wrap(kv.getRowArray(), kv.getRowOffset(), kv.getRowLength())); - kvbuilder.setFamily(HBaseZeroCopyByteString.wrap(kv.getFamilyArray(), + kvbuilder.setFamily(ByteStringer.wrap(kv.getFamilyArray(), kv.getFamilyOffset(), kv.getFamilyLength())); - kvbuilder.setQualifier(HBaseZeroCopyByteString.wrap(kv.getQualifierArray(), + kvbuilder.setQualifier(ByteStringer.wrap(kv.getQualifierArray(), kv.getQualifierOffset(), kv.getQualifierLength())); kvbuilder.setCellType(CellProtos.CellType.valueOf(kv.getTypeByte())); kvbuilder.setTimestamp(kv.getTimestamp()); - kvbuilder.setValue(HBaseZeroCopyByteString.wrap(kv.getValueArray(), kv.getValueOffset(), + kvbuilder.setValue(ByteStringer.wrap(kv.getValueArray(), kv.getValueOffset(), kv.getValueLength())); return kvbuilder.build(); } @@ -2499,9 +2499,9 @@ public final class ProtobufUtil { // input / output paths are relative to the store dir // store dir is relative to region dir CompactionDescriptor.Builder builder = CompactionDescriptor.newBuilder() - .setTableName(HBaseZeroCopyByteString.wrap(info.getTableName())) - .setEncodedRegionName(HBaseZeroCopyByteString.wrap(info.getEncodedNameAsBytes())) - .setFamilyName(HBaseZeroCopyByteString.wrap(family)) + .setTableName(ByteStringer.wrap(info.getTableName())) + .setEncodedRegionName(ByteStringer.wrap(info.getEncodedNameAsBytes())) + .setFamilyName(ByteStringer.wrap(family)) .setStoreHomeDir(storeDir.getName()); //make relative for (Path inputPath : inputPaths) { builder.addCompactionInput(inputPath.getName()); //relative path @@ -2578,8 +2578,8 @@ public final class ProtobufUtil { public static HBaseProtos.TableName toProtoTableName(TableName tableName) { return HBaseProtos.TableName.newBuilder() - .setNamespace(HBaseZeroCopyByteString.wrap(tableName.getNamespace())) - .setQualifier(HBaseZeroCopyByteString.wrap(tableName.getQualifier())).build(); + .setNamespace(ByteStringer.wrap(tableName.getNamespace())) + .setQualifier(ByteStringer.wrap(tableName.getQualifier())).build(); } public static TableName[] getTableNameArray(List tableNamesList) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java index 1a6b42d..544af85 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.protobuf; import java.io.IOException; import java.util.List; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.configuration.Configuration; import org.apache.hadoop.classification.InterfaceAudience; @@ -137,10 +137,10 @@ public final class RequestConverter { builder.setRegion(region); Column.Builder columnBuilder = Column.newBuilder(); - columnBuilder.setFamily(HBaseZeroCopyByteString.wrap(family)); + columnBuilder.setFamily(ByteStringer.wrap(family)); ClientProtos.Get.Builder getBuilder = ClientProtos.Get.newBuilder(); - getBuilder.setRow(HBaseZeroCopyByteString.wrap(row)); + getBuilder.setRow(ByteStringer.wrap(row)); getBuilder.addColumn(columnBuilder.build()); getBuilder.setClosestRowBefore(true); builder.setGet(getBuilder.build()); @@ -185,14 +185,14 @@ public final class RequestConverter { builder.setRegion(region); MutationProto.Builder mutateBuilder = MutationProto.newBuilder(); - mutateBuilder.setRow(HBaseZeroCopyByteString.wrap(row)); + mutateBuilder.setRow(ByteStringer.wrap(row)); mutateBuilder.setMutateType(MutationType.INCREMENT); mutateBuilder.setDurability(ProtobufUtil.toDurability(durability)); ColumnValue.Builder columnBuilder = ColumnValue.newBuilder(); - columnBuilder.setFamily(HBaseZeroCopyByteString.wrap(family)); + columnBuilder.setFamily(ByteStringer.wrap(family)); QualifierValue.Builder valueBuilder = QualifierValue.newBuilder(); - valueBuilder.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(amount))); - valueBuilder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + valueBuilder.setValue(ByteStringer.wrap(Bytes.toBytes(amount))); + valueBuilder.setQualifier(ByteStringer.wrap(qualifier)); columnBuilder.addQualifierValue(valueBuilder.build()); mutateBuilder.addColumnValue(columnBuilder.build()); if (nonce != HConstants.NO_NONCE) { @@ -495,7 +495,7 @@ public final class RequestConverter { builder.setRegion(region); FamilyPath.Builder familyPathBuilder = FamilyPath.newBuilder(); for (Pair familyPath: familyPaths) { - familyPathBuilder.setFamily(HBaseZeroCopyByteString.wrap(familyPath.getFirst())); + familyPathBuilder.setFamily(ByteStringer.wrap(familyPath.getFirst())); familyPathBuilder.setPath(familyPath.getSecond()); builder.addFamilyPath(familyPathBuilder.build()); } @@ -540,7 +540,7 @@ public final class RequestConverter { RegionCoprocessorServiceExec exec = (RegionCoprocessorServiceExec) row; regionActionBuilder.addAction(actionBuilder.setServiceCall( ClientProtos.CoprocessorServiceCall.newBuilder() - .setRow(HBaseZeroCopyByteString.wrap(exec.getRow())) + .setRow(ByteStringer.wrap(exec.getRow())) .setServiceName(exec.getMethod().getService().getFullName()) .setMethodName(exec.getMethod().getName()) .setRequest(exec.getRequest().toByteString()))); @@ -671,7 +671,7 @@ public final class RequestConverter { RegionSpecifier region = buildRegionSpecifier( RegionSpecifierType.REGION_NAME, regionName); builder.setRegion(region); - builder.addFamily(HBaseZeroCopyByteString.wrap(family)); + builder.addFamily(ByteStringer.wrap(family)); return builder.build(); } @@ -834,7 +834,7 @@ public final class RequestConverter { RegionSpecifierType.REGION_NAME, regionName); builder.setRegion(region); if (splitPoint != null) { - builder.setSplitPoint(HBaseZeroCopyByteString.wrap(splitPoint)); + builder.setSplitPoint(ByteStringer.wrap(splitPoint)); } return builder.build(); } @@ -874,7 +874,7 @@ public final class RequestConverter { builder.setRegion(region); builder.setMajor(major); if (family != null) { - builder.setFamily(HBaseZeroCopyByteString.wrap(family)); + builder.setFamily(ByteStringer.wrap(family)); } return builder.build(); } @@ -933,7 +933,7 @@ public final class RequestConverter { public static RegionSpecifier buildRegionSpecifier( final RegionSpecifierType type, final byte[] value) { RegionSpecifier.Builder regionBuilder = RegionSpecifier.newBuilder(); - regionBuilder.setValue(HBaseZeroCopyByteString.wrap(value)); + regionBuilder.setValue(ByteStringer.wrap(value)); regionBuilder.setType(type); return regionBuilder.build(); } @@ -954,9 +954,9 @@ public final class RequestConverter { final ByteArrayComparable comparator, final CompareType compareType) throws IOException { Condition.Builder builder = Condition.newBuilder(); - builder.setRow(HBaseZeroCopyByteString.wrap(row)); - builder.setFamily(HBaseZeroCopyByteString.wrap(family)); - builder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + builder.setRow(ByteStringer.wrap(row)); + builder.setFamily(ByteStringer.wrap(family)); + builder.setQualifier(ByteStringer.wrap(qualifier)); builder.setComparator(ProtobufUtil.toComparator(comparator)); builder.setCompareType(compareType); return builder.build(); @@ -988,7 +988,7 @@ public final class RequestConverter { final TableName tableName, final byte [] columnName) { DeleteColumnRequest.Builder builder = DeleteColumnRequest.newBuilder(); builder.setTableName(ProtobufUtil.toProtoTableName((tableName))); - builder.setColumnName(HBaseZeroCopyByteString.wrap(columnName)); + builder.setColumnName(ByteStringer.wrap(columnName)); return builder.build(); } @@ -1128,7 +1128,7 @@ public final class RequestConverter { builder.setTableSchema(hTableDesc.convert()); if (splitKeys != null) { for (byte [] splitKey : splitKeys) { - builder.addSplitKeys(HBaseZeroCopyByteString.wrap(splitKey)); + builder.addSplitKeys(ByteStringer.wrap(splitKey)); } } return builder.build(); @@ -1281,7 +1281,7 @@ public final class RequestConverter { public static GetLastFlushedSequenceIdRequest buildGetLastFlushedSequenceIdRequest( byte[] regionName) { return GetLastFlushedSequenceIdRequest.newBuilder().setRegionName( - HBaseZeroCopyByteString.wrap(regionName)).build(); + ByteStringer.wrap(regionName)).build(); } /** @@ -1336,10 +1336,10 @@ public final class RequestConverter { permissionBuilder.setTableName(ProtobufUtil.toProtoTableName(tableName)); if (family != null) { - permissionBuilder.setFamily(HBaseZeroCopyByteString.wrap(family)); + permissionBuilder.setFamily(ByteStringer.wrap(family)); } if (qualifier != null) { - permissionBuilder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + permissionBuilder.setQualifier(ByteStringer.wrap(qualifier)); } ret.setType(AccessControlProtos.Permission.Type.Table) .setTablePermission(permissionBuilder); @@ -1432,10 +1432,10 @@ public final class RequestConverter { permissionBuilder.setTableName(ProtobufUtil.toProtoTableName(tableName)); } if (family != null) { - permissionBuilder.setFamily(HBaseZeroCopyByteString.wrap(family)); + permissionBuilder.setFamily(ByteStringer.wrap(family)); } if (qualifier != null) { - permissionBuilder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + permissionBuilder.setQualifier(ByteStringer.wrap(qualifier)); } ret.setType(AccessControlProtos.Permission.Type.Table) .setTablePermission(permissionBuilder); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java index 2853aaa..0936ee1 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/EncryptionUtil.java @@ -26,7 +26,7 @@ import java.security.SecureRandom; import javax.crypto.spec.SecretKeySpec; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -83,15 +83,15 @@ public class EncryptionUtil { if (cipher.getIvLength() > 0) { iv = new byte[cipher.getIvLength()]; RNG.nextBytes(iv); - builder.setIv(HBaseZeroCopyByteString.wrap(iv)); + builder.setIv(ByteStringer.wrap(iv)); } byte[] keyBytes = key.getEncoded(); builder.setLength(keyBytes.length); - builder.setHash(HBaseZeroCopyByteString.wrap(Encryption.hash128(keyBytes))); + builder.setHash(ByteStringer.wrap(Encryption.hash128(keyBytes))); ByteArrayOutputStream out = new ByteArrayOutputStream(); Encryption.encryptWithSubjectKey(out, new ByteArrayInputStream(keyBytes), subject, conf, cipher, iv); - builder.setData(HBaseZeroCopyByteString.wrap(out.toByteArray())); + builder.setData(ByteStringer.wrap(out.toByteArray())); // Build and return the protobuf message out.reset(); builder.build().writeDelimitedTo(out); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java index e19c23c..4f21225 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.security.access; import java.io.IOException; import java.util.Map; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -86,10 +86,10 @@ public class AccessControlClient { permissionBuilder.setTableName(ProtobufUtil.toProtoTableName(tableName)); if (family != null) { - permissionBuilder.setFamily(HBaseZeroCopyByteString.wrap(family)); + permissionBuilder.setFamily(ByteStringer.wrap(family)); } if (qual != null) { - permissionBuilder.setQualifier(HBaseZeroCopyByteString.wrap(qual)); + permissionBuilder.setQualifier(ByteStringer.wrap(qual)); } ret.setType(AccessControlProtos.Permission.Type.Table).setTablePermission( permissionBuilder); @@ -150,10 +150,10 @@ public class AccessControlClient { permissionBuilder.setTableName(ProtobufUtil.toProtoTableName(tableName)); } if (family != null) { - permissionBuilder.setFamily(HBaseZeroCopyByteString.wrap(family)); + permissionBuilder.setFamily(ByteStringer.wrap(family)); } if (qualifier != null) { - permissionBuilder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + permissionBuilder.setQualifier(ByteStringer.wrap(qualifier)); } ret.setType(AccessControlProtos.Permission.Type.Table).setTablePermission( permissionBuilder); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java index bf1b01f..8a17994 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java @@ -22,7 +22,7 @@ import static org.apache.hadoop.hbase.security.visibility.VisibilityConstants.LA import java.io.IOException; import java.util.Map; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -86,7 +86,7 @@ public class VisibilityClient { for (String label : labels) { if (label.length() > 0) { VisibilityLabel.Builder newBuilder = VisibilityLabel.newBuilder(); - newBuilder.setLabel(HBaseZeroCopyByteString.wrap(Bytes.toBytes(label))); + newBuilder.setLabel(ByteStringer.wrap(Bytes.toBytes(label))); builder.addVisLabel(newBuilder.build()); } } @@ -137,7 +137,7 @@ public class VisibilityClient { public GetAuthsResponse call(VisibilityLabelsService service) throws IOException { GetAuthsRequest.Builder getAuthReqBuilder = GetAuthsRequest.newBuilder(); - getAuthReqBuilder.setUser(HBaseZeroCopyByteString.wrap(Bytes.toBytes(user))); + getAuthReqBuilder.setUser(ByteStringer.wrap(Bytes.toBytes(user))); service.getAuths(controller, getAuthReqBuilder.build(), rpcCallback); return rpcCallback.get(); } @@ -179,10 +179,10 @@ public class VisibilityClient { public VisibilityLabelsResponse call(VisibilityLabelsService service) throws IOException { SetAuthsRequest.Builder setAuthReqBuilder = SetAuthsRequest.newBuilder(); - setAuthReqBuilder.setUser(HBaseZeroCopyByteString.wrap(Bytes.toBytes(user))); + setAuthReqBuilder.setUser(ByteStringer.wrap(Bytes.toBytes(user))); for (String auth : auths) { if (auth.length() > 0) { - setAuthReqBuilder.addAuth(HBaseZeroCopyByteString.wrap(Bytes.toBytes(auth))); + setAuthReqBuilder.addAuth(ByteStringer.wrap(Bytes.toBytes(auth))); } } if (setOrClear) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java index 1c503e0..2a8b51a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java @@ -35,7 +35,7 @@ import java.util.Properties; import javax.security.auth.login.AppConfigurationEntry; import javax.security.auth.login.AppConfigurationEntry.LoginModuleControlFlag; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -1949,7 +1949,7 @@ public class ZKUtil { for (Map.Entry e : storeSequenceIds.entrySet()){ byte[] columnFamilyName = e.getKey(); Long curSeqId = e.getValue(); - storeSequenceIdBuilder.setFamilyName(HBaseZeroCopyByteString.wrap(columnFamilyName)); + storeSequenceIdBuilder.setFamilyName(ByteStringer.wrap(columnFamilyName)); storeSequenceIdBuilder.setSequenceId(curSeqId); regionSequenceIdsBuilder.addStoreSequenceId(storeSequenceIdBuilder.build()); storeSequenceIdBuilder.clear(); diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java index 7ee4476..fa0819d 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java @@ -33,7 +33,7 @@ import java.util.concurrent.Executors; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicLong; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.lang.NotImplementedException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -510,7 +510,7 @@ public class TestClientNoCluster extends Configured implements Tool { if (max <= 0) break; if (++count > max) break; HRegionInfo hri = e.getValue().getFirst(); - ByteString row = HBaseZeroCopyByteString.wrap(hri.getRegionName()); + ByteString row = ByteStringer.wrap(hri.getRegionName()); resultBuilder.clear(); resultBuilder.addCell(getRegionInfo(row, hri)); resultBuilder.addCell(getServer(row, e.getValue().getSecond())); @@ -565,11 +565,11 @@ public class TestClientNoCluster extends Configured implements Tool { } private final static ByteString CATALOG_FAMILY_BYTESTRING = - HBaseZeroCopyByteString.wrap(HConstants.CATALOG_FAMILY); + ByteStringer.wrap(HConstants.CATALOG_FAMILY); private final static ByteString REGIONINFO_QUALIFIER_BYTESTRING = - HBaseZeroCopyByteString.wrap(HConstants.REGIONINFO_QUALIFIER); + ByteStringer.wrap(HConstants.REGIONINFO_QUALIFIER); private final static ByteString SERVER_QUALIFIER_BYTESTRING = - HBaseZeroCopyByteString.wrap(HConstants.SERVER_QUALIFIER); + ByteStringer.wrap(HConstants.SERVER_QUALIFIER); static CellProtos.Cell.Builder getBaseCellBuilder(final ByteString row) { CellProtos.Cell.Builder cellBuilder = CellProtos.Cell.newBuilder(); @@ -582,7 +582,7 @@ public class TestClientNoCluster extends Configured implements Tool { static CellProtos.Cell getRegionInfo(final ByteString row, final HRegionInfo hri) { CellProtos.Cell.Builder cellBuilder = getBaseCellBuilder(row); cellBuilder.setQualifier(REGIONINFO_QUALIFIER_BYTESTRING); - cellBuilder.setValue(HBaseZeroCopyByteString.wrap(hri.toByteArray())); + cellBuilder.setValue(ByteStringer.wrap(hri.toByteArray())); return cellBuilder.build(); } @@ -595,9 +595,9 @@ public class TestClientNoCluster extends Configured implements Tool { static CellProtos.Cell getStartCode(final ByteString row) { CellProtos.Cell.Builder cellBuilder = getBaseCellBuilder(row); - cellBuilder.setQualifier(HBaseZeroCopyByteString.wrap(HConstants.STARTCODE_QUALIFIER)); + cellBuilder.setQualifier(ByteStringer.wrap(HConstants.STARTCODE_QUALIFIER)); // TODO: - cellBuilder.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(META_SERVERNAME.getStartcode()))); + cellBuilder.setValue(ByteStringer.wrap(Bytes.toBytes(META_SERVERNAME.getStartcode()))); return cellBuilder.build(); } diff --git a/hbase-common/pom.xml b/hbase-common/pom.xml index 7dbcd93..893aed0 100644 --- a/hbase-common/pom.xml +++ b/hbase-common/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-examples/pom.xml b/hbase-examples/pom.xml index a00d382..82ded72 100644 --- a/hbase-examples/pom.xml +++ b/hbase-examples/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. hbase-examples diff --git a/hbase-hadoop-compat/pom.xml b/hbase-hadoop-compat/pom.xml index 0402305..97be5af 100644 --- a/hbase-hadoop-compat/pom.xml +++ b/hbase-hadoop-compat/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-hadoop1-compat/pom.xml b/hbase-hadoop1-compat/pom.xml index 81ab14e..35d95b6 100644 --- a/hbase-hadoop1-compat/pom.xml +++ b/hbase-hadoop1-compat/pom.xml @@ -21,7 +21,7 @@ limitations under the License. hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-hadoop2-compat/pom.xml b/hbase-hadoop2-compat/pom.xml index 3da5e5f..78a166a 100644 --- a/hbase-hadoop2-compat/pom.xml +++ b/hbase-hadoop2-compat/pom.xml @@ -21,7 +21,7 @@ limitations under the License. hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-it/pom.xml b/hbase-it/pom.xml index 650f890..9424cf0 100644 --- a/hbase-it/pom.xml +++ b/hbase-it/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-prefix-tree/pom.xml b/hbase-prefix-tree/pom.xml index 1cc5f25..bc70a56 100644 --- a/hbase-prefix-tree/pom.xml +++ b/hbase-prefix-tree/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. diff --git a/hbase-protocol/pom.xml b/hbase-protocol/pom.xml index fc3a833..c699255 100644 --- a/hbase-protocol/pom.xml +++ b/hbase-protocol/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. @@ -108,6 +108,10 @@ com.google.protobuf protobuf-java + + commons-logging + commons-logging + diff --git a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java index 68c4f6f..f7e9c7f 100644 --- a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java +++ b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java @@ -64,5 +64,4 @@ public final class HBaseZeroCopyByteString extends LiteralByteString { throw new UnsupportedOperationException("Need a LiteralByteString, got a " + buf.getClass().getName()); } - -} +} \ No newline at end of file diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/util/ByteStringer.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/util/ByteStringer.java new file mode 100644 index 0000000..4656e83 --- /dev/null +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/util/ByteStringer.java @@ -0,0 +1,68 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.util; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; + +import com.google.protobuf.ByteString; +import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; + +/** + * Hack to workaround HBASE-1304 issue that keeps bubbling up when a mapreduce context. + */ +public class ByteStringer { + private static final Log LOG = LogFactory.getLog(ByteStringer.class); + + /** + * Flag set at class loading time. + */ + private static boolean USE_ZEROCOPYBYTESTRING = true; + + // Can I classload HBaseZeroCopyByteString without IllegalAccessError? + // If we can, use it passing ByteStrings to pb else use native ByteString though more costly + // because it makes a copy of the passed in array. + static { + try { + HBaseZeroCopyByteString.wrap(new byte [0]); + } catch (IllegalAccessError iae) { + USE_ZEROCOPYBYTESTRING = false; + LOG.debug("Failed to classload HBaseZeroCopyByteString: " + iae.toString()); + } + } + + private ByteStringer() { + super(); + } + + /** + * Wraps a byte array in a {@link ByteString} without copying it. + */ + public static ByteString wrap(final byte[] array) { + return USE_ZEROCOPYBYTESTRING? HBaseZeroCopyByteString.wrap(array): ByteString.copyFrom(array); + } + + /** + * Wraps a subset of a byte array in a {@link ByteString} without copying it. + */ + public static ByteString wrap(final byte[] array, int offset, int length) { + return USE_ZEROCOPYBYTESTRING? ByteStringer.wrap(array, offset, length): + ByteString.copyFrom(array, offset, length); + } +} diff --git a/hbase-server/pom.xml b/hbase-server/pom.xml index 5f4eee6..3dd4940 100644 --- a/hbase-server/pom.xml +++ b/hbase-server/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. hbase-server diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java index db6bea8..00b713c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/codec/MessageCodec.java @@ -21,7 +21,7 @@ import java.io.IOException; import java.io.InputStream; import java.io.OutputStream; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; @@ -44,15 +44,15 @@ public class MessageCodec implements Codec { CellProtos.Cell.Builder builder = CellProtos.Cell.newBuilder(); // This copies bytes from Cell to ByteString. I don't see anyway around the copy. // ByteString is final. - builder.setRow(HBaseZeroCopyByteString.wrap(cell.getRowArray(), cell.getRowOffset(), + builder.setRow(ByteStringer.wrap(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength())); - builder.setFamily(HBaseZeroCopyByteString.wrap(cell.getFamilyArray(), cell.getFamilyOffset(), + builder.setFamily(ByteStringer.wrap(cell.getFamilyArray(), cell.getFamilyOffset(), cell.getFamilyLength())); - builder.setQualifier(HBaseZeroCopyByteString.wrap(cell.getQualifierArray(), + builder.setQualifier(ByteStringer.wrap(cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength())); builder.setTimestamp(cell.getTimestamp()); builder.setCellType(CellProtos.CellType.valueOf(cell.getTypeByte())); - builder.setValue(HBaseZeroCopyByteString.wrap(cell.getValueArray(), cell.getValueOffset(), + builder.setValue(ByteStringer.wrap(cell.getValueArray(), cell.getValueOffset(), cell.getValueLength())); CellProtos.Cell pbcell = builder.build(); pbcell.writeDelimitedTo(this.out); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java index 7ed9f6b..de05575 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/Reference.java @@ -24,7 +24,7 @@ import java.io.DataInputStream; import java.io.IOException; import java.io.InputStream; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -193,7 +193,7 @@ public class Reference { FSProtos.Reference.Builder builder = FSProtos.Reference.newBuilder(); builder.setRange(isTopFileRegion(getFileRegion())? FSProtos.Reference.Range.TOP: FSProtos.Reference.Range.BOTTOM); - builder.setSplitkey(HBaseZeroCopyByteString.wrap(getSplitKey())); + builder.setSplitkey(ByteStringer.wrap(getSplitKey())); return builder.build(); } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java index 8428493..1b065ec 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java @@ -26,7 +26,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.hbase.KeyValue; @@ -205,7 +205,7 @@ public class FixedFileTrailer { .setComparatorClassName(comparatorClassName) .setCompressionCodec(compressionCodec.ordinal()); if (encryptionKey != null) { - builder.setEncryptionKey(HBaseZeroCopyByteString.wrap(encryptionKey)); + builder.setEncryptionKey(ByteStringer.wrap(encryptionKey)); } // We need this extra copy unfortunately to determine the final size of the // delimited output, see use of baos.size() below. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java index 71f380a..4d1c2fd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java @@ -40,7 +40,7 @@ import java.util.concurrent.BlockingQueue; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicLong; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -739,8 +739,8 @@ public class HFile { HFileProtos.FileInfoProto.Builder builder = HFileProtos.FileInfoProto.newBuilder(); for (Map.Entry e: this.map.entrySet()) { HBaseProtos.BytesBytesPair.Builder bbpBuilder = HBaseProtos.BytesBytesPair.newBuilder(); - bbpBuilder.setFirst(HBaseZeroCopyByteString.wrap(e.getKey())); - bbpBuilder.setSecond(HBaseZeroCopyByteString.wrap(e.getValue())); + bbpBuilder.setFirst(ByteStringer.wrap(e.getKey())); + bbpBuilder.setSecond(ByteStringer.wrap(e.getValue())); builder.addMapEntry(bbpBuilder.build()); } out.write(ProtobufUtil.PB_MAGIC); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java index 89702c5..c43b036 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java @@ -1220,7 +1220,6 @@ public class HFileBlock implements Cacheable { } else { // Positional read. Better for random reads; or when the streamLock is already locked. int extraSize = peekIntoNextBlock ? hdrSize : 0; - int ret = istream.read(fileOffset, dest, destOffset, size + extraSize); if (ret < size) { throw new IOException("Positional read of " + size + " bytes " + @@ -1234,8 +1233,7 @@ public class HFileBlock implements Cacheable { } assert peekIntoNextBlock; - return Bytes.toInt(dest, destOffset + size + BlockType.MAGIC_LENGTH) + - hdrSize; + return Bytes.toInt(dest, destOffset + size + BlockType.MAGIC_LENGTH) + hdrSize; } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java index d156b71..1ce10ca 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java @@ -28,7 +28,7 @@ import java.util.Map; import java.util.NavigableMap; import java.util.UUID; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellScanner; @@ -91,8 +91,8 @@ public class ReplicationProtbufUtil { WALProtos.WALKey.Builder keyBuilder = entryBuilder.getKeyBuilder(); HLogKey key = entry.getKey(); keyBuilder.setEncodedRegionName( - HBaseZeroCopyByteString.wrap(key.getEncodedRegionName())); - keyBuilder.setTableName(HBaseZeroCopyByteString.wrap(key.getTablename().getName())); + ByteStringer.wrap(key.getEncodedRegionName())); + keyBuilder.setTableName(ByteStringer.wrap(key.getTablename().getName())); keyBuilder.setLogSequenceNumber(key.getLogSeqNum()); keyBuilder.setWriteTime(key.getWriteTime()); if (key.getNonce() != HConstants.NO_NONCE) { @@ -110,7 +110,7 @@ public class ReplicationProtbufUtil { NavigableMap scopes = key.getScopes(); if (scopes != null && !scopes.isEmpty()) { for (Map.Entry scope: scopes.entrySet()) { - scopeBuilder.setFamily(HBaseZeroCopyByteString.wrap(scope.getKey())); + scopeBuilder.setFamily(ByteStringer.wrap(scope.getKey())); WALProtos.ScopeType scopeType = WALProtos.ScopeType.valueOf(scope.getValue().intValue()); scopeBuilder.setScopeType(scopeType); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java index 541463e..edc190f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java @@ -52,7 +52,7 @@ import java.util.concurrent.locks.ReentrantReadWriteLock; import javax.management.ObjectName; import com.google.common.annotations.VisibleForTesting; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -1349,7 +1349,7 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa regionSpecifier = RegionSpecifier.newBuilder(); } regionSpecifier.setType(RegionSpecifierType.REGION_NAME); - regionSpecifier.setValue(HBaseZeroCopyByteString.wrap(name)); + regionSpecifier.setValue(ByteStringer.wrap(name)); regionLoadBldr.setRegionSpecifier(regionSpecifier.build()) .setStores(stores) .setStorefiles(storefiles) @@ -4061,7 +4061,7 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa RollWALWriterResponse.Builder builder = RollWALWriterResponse.newBuilder(); if (regionsToFlush != null) { for (byte[] region: regionsToFlush) { - builder.addRegionToFlush(HBaseZeroCopyByteString.wrap(region)); + builder.addRegionToFlush(ByteStringer.wrap(region)); } } return builder.build(); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java index 42c3f95..b5d83f6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java @@ -31,7 +31,7 @@ import java.util.NavigableMap; import java.util.TreeMap; import java.util.UUID; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -443,8 +443,8 @@ public class HLogKey implements WritableComparable { WALCellCodec.ByteStringCompressor compressor) throws IOException { WALKey.Builder builder = WALKey.newBuilder(); if (compressionContext == null) { - builder.setEncodedRegionName(HBaseZeroCopyByteString.wrap(this.encodedRegionName)); - builder.setTableName(HBaseZeroCopyByteString.wrap(this.tablename.getName())); + builder.setEncodedRegionName(ByteStringer.wrap(this.encodedRegionName)); + builder.setTableName(ByteStringer.wrap(this.tablename.getName())); } else { builder.setEncodedRegionName( compressor.compress(this.encodedRegionName, compressionContext.regionDict)); @@ -467,7 +467,7 @@ public class HLogKey implements WritableComparable { } if (scopes != null) { for (Map.Entry e : scopes.entrySet()) { - ByteString family = (compressionContext == null) ? HBaseZeroCopyByteString.wrap(e.getKey()) + ByteString family = (compressionContext == null) ? ByteStringer.wrap(e.getKey()) : compressor.compress(e.getKey(), compressionContext.familyDict); builder.addScopes(FamilyScope.newBuilder() .setFamily(family).setScopeType(ScopeType.valueOf(e.getValue()))); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogWriter.java index 8a09af4..ee100fd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogWriter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogWriter.java @@ -24,7 +24,7 @@ import java.security.SecureRandom; import javax.crypto.spec.SecretKeySpec; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.HConstants; @@ -57,7 +57,7 @@ public class SecureProtobufLogWriter extends ProtobufLogWriter { byte[] keyBytes = new byte[cipher.getKeyLength()]; rng.nextBytes(keyBytes); Key key = new SecretKeySpec(keyBytes, cipher.getName()); - builder.setEncryptionKey(HBaseZeroCopyByteString.wrap(EncryptionUtil.wrapKey(conf, + builder.setEncryptionKey(ByteStringer.wrap(EncryptionUtil.wrapKey(conf, conf.get(HConstants.CRYPTO_WAL_KEY_NAME_CONF_KEY, conf.get(HConstants.CRYPTO_MASTERKEY_NAME_CONF_KEY, User.getCurrent().getShortName())), diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java index c01359d..66fb85a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java @@ -28,7 +28,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.bind.annotation.XmlValue; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; @@ -184,8 +184,8 @@ public class CellModel implements ProtobufMessageHandler, Serializable { @Override public byte[] createProtobufOutput() { Cell.Builder builder = Cell.newBuilder(); - builder.setColumn(HBaseZeroCopyByteString.wrap(getColumn())); - builder.setData(HBaseZeroCopyByteString.wrap(getValue())); + builder.setColumn(ByteStringer.wrap(getColumn())); + builder.setData(ByteStringer.wrap(getValue())); if (hasUserTimestamp()) { builder.setTimestamp(getTimestamp()); } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java index 5261484..143c1d6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java @@ -29,7 +29,7 @@ import javax.xml.bind.annotation.XmlAccessorType; import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; @@ -114,11 +114,11 @@ public class CellSetModel implements Serializable, ProtobufMessageHandler { CellSet.Builder builder = CellSet.newBuilder(); for (RowModel row: getRows()) { CellSet.Row.Builder rowBuilder = CellSet.Row.newBuilder(); - rowBuilder.setKey(HBaseZeroCopyByteString.wrap(row.getKey())); + rowBuilder.setKey(ByteStringer.wrap(row.getKey())); for (CellModel cell: row.getCells()) { Cell.Builder cellBuilder = Cell.newBuilder(); - cellBuilder.setColumn(HBaseZeroCopyByteString.wrap(cell.getColumn())); - cellBuilder.setData(HBaseZeroCopyByteString.wrap(cell.getValue())); + cellBuilder.setColumn(ByteStringer.wrap(cell.getColumn())); + cellBuilder.setData(ByteStringer.wrap(cell.getValue())); if (cell.hasUserTimestamp()) { cellBuilder.setTimestamp(cell.getTimestamp()); } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java index 7950f55..2fa53f4 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java @@ -75,7 +75,7 @@ import org.apache.hadoop.hbase.util.Base64; import org.apache.hadoop.hbase.util.Bytes; import com.google.protobuf.ByteString; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.sun.jersey.api.json.JSONConfiguration; import com.sun.jersey.api.json.JSONJAXBContext; import com.sun.jersey.api.json.JSONMarshaller; @@ -781,13 +781,13 @@ public class ScannerModel implements ProtobufMessageHandler, Serializable { public byte[] createProtobufOutput() { Scanner.Builder builder = Scanner.newBuilder(); if (!Bytes.equals(startRow, HConstants.EMPTY_START_ROW)) { - builder.setStartRow(HBaseZeroCopyByteString.wrap(startRow)); + builder.setStartRow(ByteStringer.wrap(startRow)); } if (!Bytes.equals(endRow, HConstants.EMPTY_START_ROW)) { - builder.setEndRow(HBaseZeroCopyByteString.wrap(endRow)); + builder.setEndRow(ByteStringer.wrap(endRow)); } for (byte[] column: columns) { - builder.addColumns(HBaseZeroCopyByteString.wrap(column)); + builder.addColumns(ByteStringer.wrap(column)); } if (startTime != 0) { builder.setStartTime(startTime); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java index 3c77925..dd9f04e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java @@ -29,7 +29,7 @@ import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlElementWrapper; import javax.xml.bind.annotation.XmlRootElement; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus; @@ -721,7 +721,7 @@ public class StorageClusterStatusModel for (Node.Region region: node.regions) { StorageClusterStatus.Region.Builder regionBuilder = StorageClusterStatus.Region.newBuilder(); - regionBuilder.setName(HBaseZeroCopyByteString.wrap(region.name)); + regionBuilder.setName(ByteStringer.wrap(region.name)); regionBuilder.setStores(region.stores); regionBuilder.setStorefiles(region.storefiles); regionBuilder.setStorefileSizeMB(region.storefileSizeMB); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java index 63502e1..0847be9 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java @@ -28,7 +28,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo; @@ -134,8 +134,8 @@ public class TableInfoModel implements Serializable, ProtobufMessageHandler { TableInfo.Region.Builder regionBuilder = TableInfo.Region.newBuilder(); regionBuilder.setName(aRegion.getName()); regionBuilder.setId(aRegion.getId()); - regionBuilder.setStartKey(HBaseZeroCopyByteString.wrap(aRegion.getStartKey())); - regionBuilder.setEndKey(HBaseZeroCopyByteString.wrap(aRegion.getEndKey())); + regionBuilder.setStartKey(ByteStringer.wrap(aRegion.getStartKey())); + regionBuilder.setEndKey(ByteStringer.wrap(aRegion.getEndKey())); regionBuilder.setLocation(aRegion.getLocation()); builder.addRegions(regionBuilder); } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java index e634840..2ce0a66 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java @@ -116,7 +116,7 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import com.google.common.collect.Lists; import com.google.common.collect.MapMaker; import com.google.protobuf.ByteString; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import com.google.protobuf.RpcCallback; import com.google.protobuf.RpcController; import com.google.protobuf.Service; @@ -1280,7 +1280,7 @@ public class VisibilityController extends BaseRegionObserver implements MasterOb try { List labels = getUserAuthsFromLabelsTable(user); for (String label : labels) { - response.addAuth(HBaseZeroCopyByteString.wrap(Bytes.toBytes(label))); + response.addAuth(ByteStringer.wrap(Bytes.toBytes(label))); } } catch (IOException e) { ResponseConverter.setControllerException(controller, e); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java index 0c7764f..b020710 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java @@ -23,7 +23,7 @@ import java.util.List; import java.util.Map; import java.util.Map.Entry; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.lang.StringUtils; import org.apache.hadoop.classification.InterfaceAudience; @@ -61,7 +61,7 @@ public class VisibilityUtils { VisibilityLabelsRequest.Builder visReqBuilder = VisibilityLabelsRequest.newBuilder(); for (Entry entry : existingLabels.entrySet()) { VisibilityLabel.Builder visLabBuilder = VisibilityLabel.newBuilder(); - visLabBuilder.setLabel(HBaseZeroCopyByteString.wrap(Bytes.toBytes(entry.getKey()))); + visLabBuilder.setLabel(ByteStringer.wrap(Bytes.toBytes(entry.getKey()))); visLabBuilder.setOrdinal(entry.getValue()); visReqBuilder.addVisLabel(visLabBuilder.build()); } @@ -77,7 +77,7 @@ public class VisibilityUtils { MultiUserAuthorizations.Builder builder = MultiUserAuthorizations.newBuilder(); for (Entry> entry : userAuths.entrySet()) { UserAuthorizations.Builder userAuthsBuilder = UserAuthorizations.newBuilder(); - userAuthsBuilder.setUser(HBaseZeroCopyByteString.wrap(Bytes.toBytes(entry.getKey()))); + userAuthsBuilder.setUser(ByteStringer.wrap(Bytes.toBytes(entry.getKey()))); for (Integer label : entry.getValue()) { userAuthsBuilder.addAuth(label); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java index be5f79d..ed0eb97 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java @@ -26,7 +26,7 @@ import java.util.Collections; import java.util.Map; import java.util.TreeMap; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -111,9 +111,9 @@ public class TestBatchCoprocessorEndpoint { ColumnAggregationWithNullResponseProtos.SumRequest.Builder builder = ColumnAggregationWithNullResponseProtos.SumRequest .newBuilder(); - builder.setFamily(HBaseZeroCopyByteString.wrap(TEST_FAMILY)); + builder.setFamily(ByteStringer.wrap(TEST_FAMILY)); if (TEST_QUALIFIER != null && TEST_QUALIFIER.length > 0) { - builder.setQualifier(HBaseZeroCopyByteString.wrap(TEST_QUALIFIER)); + builder.setQualifier(ByteStringer.wrap(TEST_QUALIFIER)); } Map results = table.batchCoprocessorService( @@ -149,9 +149,9 @@ public class TestBatchCoprocessorEndpoint { Throwable { ColumnAggregationProtos.SumRequest.Builder builder = ColumnAggregationProtos.SumRequest .newBuilder(); - builder.setFamily(HBaseZeroCopyByteString.wrap(family)); + builder.setFamily(ByteStringer.wrap(family)); if (qualifier != null && qualifier.length > 0) { - builder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + builder.setQualifier(ByteStringer.wrap(qualifier)); } return table.batchCoprocessorService( ColumnAggregationProtos.ColumnAggregationService.getDescriptor().findMethodByName("sum"), @@ -238,9 +238,9 @@ public class TestBatchCoprocessorEndpoint { ColumnAggregationWithErrorsProtos.SumRequest.Builder builder = ColumnAggregationWithErrorsProtos.SumRequest .newBuilder(); - builder.setFamily(HBaseZeroCopyByteString.wrap(TEST_FAMILY)); + builder.setFamily(ByteStringer.wrap(TEST_FAMILY)); if (TEST_QUALIFIER != null && TEST_QUALIFIER.length > 0) { - builder.setQualifier(HBaseZeroCopyByteString.wrap(TEST_QUALIFIER)); + builder.setQualifier(ByteStringer.wrap(TEST_QUALIFIER)); } boolean hasError = false; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java index 3cc73c1..c121c57 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java @@ -30,7 +30,7 @@ import java.util.Map; import java.util.NavigableMap; import java.util.TreeMap; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -125,9 +125,9 @@ public class TestCoprocessorEndpoint { new BlockingRpcCallback(); ColumnAggregationProtos.SumRequest.Builder builder = ColumnAggregationProtos.SumRequest.newBuilder(); - builder.setFamily(HBaseZeroCopyByteString.wrap(family)); + builder.setFamily(ByteStringer.wrap(family)); if (qualifier != null && qualifier.length > 0) { - builder.setQualifier(HBaseZeroCopyByteString.wrap(qualifier)); + builder.setQualifier(ByteStringer.wrap(qualifier)); } instance.sum(null, builder.build(), rpcCallback); return rpcCallback.get().getSum(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java index 28db1aa..f6573f5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java @@ -33,7 +33,7 @@ import java.util.concurrent.CountDownLatch; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicLong; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; @@ -362,7 +362,7 @@ public class TestRowProcessorEndpoint { public IncCounterProcessorRequest getRequestData() throws IOException { IncCounterProcessorRequest.Builder builder = IncCounterProcessorRequest.newBuilder(); builder.setCounter(counter); - builder.setRow(HBaseZeroCopyByteString.wrap(row)); + builder.setRow(ByteStringer.wrap(row)); return builder.build(); } @@ -441,8 +441,8 @@ public class TestRowProcessorEndpoint { public FriendsOfFriendsProcessorRequest getRequestData() throws IOException { FriendsOfFriendsProcessorRequest.Builder builder = FriendsOfFriendsProcessorRequest.newBuilder(); - builder.setPerson(HBaseZeroCopyByteString.wrap(person)); - builder.setRow(HBaseZeroCopyByteString.wrap(row)); + builder.setPerson(ByteStringer.wrap(person)); + builder.setRow(ByteStringer.wrap(row)); builder.addAllResult(result); FriendsOfFriendsProcessorRequest f = builder.build(); return f; @@ -546,8 +546,8 @@ public class TestRowProcessorEndpoint { @Override public RowSwapProcessorRequest getRequestData() throws IOException { RowSwapProcessorRequest.Builder builder = RowSwapProcessorRequest.newBuilder(); - builder.setRow1(HBaseZeroCopyByteString.wrap(row1)); - builder.setRow2(HBaseZeroCopyByteString.wrap(row2)); + builder.setRow1(ByteStringer.wrap(row1)); + builder.setRow2(ByteStringer.wrap(row2)); return builder.build(); } @@ -606,7 +606,7 @@ public class TestRowProcessorEndpoint { @Override public TimeoutProcessorRequest getRequestData() throws IOException { TimeoutProcessorRequest.Builder builder = TimeoutProcessorRequest.newBuilder(); - builder.setRow(HBaseZeroCopyByteString.wrap(row)); + builder.setRow(ByteStringer.wrap(row)); return builder.build(); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java index 0aa011b..f4b71f7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/protobuf/TestProtobufUtil.java @@ -22,7 +22,7 @@ import static org.junit.Assert.assertEquals; import java.io.IOException; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.SmallTests; import org.apache.hadoop.hbase.client.Append; @@ -55,12 +55,12 @@ public class TestProtobufUtil { NameBytesPair.Builder builder = NameBytesPair.newBuilder(); final String omg = "OMG!!!"; builder.setName("java.io.IOException"); - builder.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(omg))); + builder.setValue(ByteStringer.wrap(Bytes.toBytes(omg))); Throwable t = ProtobufUtil.toException(builder.build()); assertEquals(omg, t.getMessage()); builder.clear(); builder.setName("org.apache.hadoop.ipc.RemoteException"); - builder.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(omg))); + builder.setValue(ByteStringer.wrap(Bytes.toBytes(omg))); t = ProtobufUtil.toException(builder.build()); assertEquals(omg, t.getMessage()); } @@ -200,10 +200,10 @@ public class TestProtobufUtil { valueBuilder.setFamily(ByteString.copyFromUtf8("f1")); QualifierValue.Builder qualifierBuilder = QualifierValue.newBuilder(); qualifierBuilder.setQualifier(ByteString.copyFromUtf8("c1")); - qualifierBuilder.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(11L))); + qualifierBuilder.setValue(ByteStringer.wrap(Bytes.toBytes(11L))); valueBuilder.addQualifierValue(qualifierBuilder.build()); qualifierBuilder.setQualifier(ByteString.copyFromUtf8("c2")); - qualifierBuilder.setValue(HBaseZeroCopyByteString.wrap(Bytes.toBytes(22L))); + qualifierBuilder.setValue(ByteStringer.wrap(Bytes.toBytes(22L))); valueBuilder.addQualifierValue(qualifierBuilder.build()); mutateBuilder.addColumnValue(valueBuilder.build()); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java index 4aa11b5..55b01a4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestPriorityRpc.java @@ -24,7 +24,7 @@ import static org.junit.Assert.assertTrue; import java.io.IOException; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -71,12 +71,12 @@ public class TestPriorityRpc { GetRequest.Builder getRequestBuilder = GetRequest.newBuilder(); RegionSpecifier.Builder regionSpecifierBuilder = RegionSpecifier.newBuilder(); regionSpecifierBuilder.setType(RegionSpecifierType.REGION_NAME); - ByteString name = HBaseZeroCopyByteString.wrap(HRegionInfo.FIRST_META_REGIONINFO.getRegionName()); + ByteString name = ByteStringer.wrap(HRegionInfo.FIRST_META_REGIONINFO.getRegionName()); regionSpecifierBuilder.setValue(name); RegionSpecifier regionSpecifier = regionSpecifierBuilder.build(); getRequestBuilder.setRegion(regionSpecifier); Get.Builder getBuilder = Get.newBuilder(); - getBuilder.setRow(HBaseZeroCopyByteString.wrap("somerow".getBytes())); + getBuilder.setRow(ByteStringer.wrap("somerow".getBytes())); getRequestBuilder.setGet(getBuilder.build()); GetRequest getRequest = getRequestBuilder.build(); RequestHeader header = headerBuilder.build(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java index 2a08ef2..3a138e4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java @@ -25,7 +25,7 @@ import java.util.ArrayList; import java.util.List; import java.util.concurrent.atomic.AtomicBoolean; -import com.google.protobuf.HBaseZeroCopyByteString; +import org.apache.hadoop.hbase.util.ByteStringer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -261,9 +261,9 @@ public class TestReplicationSink { uuidBuilder.setLeastSigBits(HConstants.DEFAULT_CLUSTER_ID.getLeastSignificantBits()); uuidBuilder.setMostSigBits(HConstants.DEFAULT_CLUSTER_ID.getMostSignificantBits()); keyBuilder.setClusterId(uuidBuilder.build()); - keyBuilder.setTableName(HBaseZeroCopyByteString.wrap(table)); + keyBuilder.setTableName(ByteStringer.wrap(table)); keyBuilder.setWriteTime(now); - keyBuilder.setEncodedRegionName(HBaseZeroCopyByteString.wrap(HConstants.EMPTY_BYTE_ARRAY)); + keyBuilder.setEncodedRegionName(ByteStringer.wrap(HConstants.EMPTY_BYTE_ARRAY)); keyBuilder.setLogSequenceNumber(-1); builder.setKey(keyBuilder.build()); cells.add(kv); diff --git a/hbase-shell/pom.xml b/hbase-shell/pom.xml index cac060d..779bb41 100644 --- a/hbase-shell/pom.xml +++ b/hbase-shell/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. hbase-shell diff --git a/hbase-testing-util/pom.xml b/hbase-testing-util/pom.xml index 81ba938..7cb7d7c 100644 --- a/hbase-testing-util/pom.xml +++ b/hbase-testing-util/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. hbase-testing-util diff --git a/hbase-thrift/pom.xml b/hbase-thrift/pom.xml index 5ce162b..32570b7 100644 --- a/hbase-thrift/pom.xml +++ b/hbase-thrift/pom.xml @@ -23,7 +23,7 @@ hbase org.apache.hbase - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT .. hbase-thrift diff --git a/pom.xml b/pom.xml index fbfa819..d4f24fe 100644 --- a/pom.xml +++ b/pom.xml @@ -39,7 +39,7 @@ org.apache.hbase hbase pom - 0.98.4-SNAPSHOT + 0.98.4-11118-SNAPSHOT HBase Apache HBase™ is the Hadoop database. Use it when you need