diff --git hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java index 2fbfd9f..d3d7608 100644 --- hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java +++ hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java @@ -1713,7 +1713,7 @@ public class MetaTableAccessor { public static Put makePutFromTableState(TableState state) { long time = EnvironmentEdgeManager.currentTime(); Put put = new Put(state.getTableName().getName(), time); - put.add(getTableFamily(), getStateColumn(), state.convert().toByteArray()); + put.addColumn(getTableFamily(), getStateColumn(), state.convert().toByteArray()); return put; } diff --git hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java index 5e0a341..25fa5fd 100644 --- hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java +++ hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java @@ -137,26 +137,13 @@ public class Put extends Mutation implements HeapSize, Comparable { * @param qualifier column qualifier * @param value column value * @return this - * @deprecated Since 1.0.0. Use {@link #addColumn(byte[], byte[], byte[])} - */ - @Deprecated - public Put add(byte [] family, byte [] qualifier, byte [] value) { - return addColumn(family, qualifier, value); - } - - /** - * Add the specified column and value to this Put operation. - * @param family family name - * @param qualifier column qualifier - * @param value column value - * @return this */ public Put addColumn(byte [] family, byte [] qualifier, byte [] value) { return addColumn(family, qualifier, this.ts, value); } /** - * See {@link #add(byte[], byte[], byte[])}. This version expects + * See {@link #addColumn(byte[], byte[], byte[])}. This version expects * that the underlying arrays won't change. It's intended * for usage internal HBase to and for advanced client applications. */ @@ -183,21 +170,6 @@ public class Put extends Mutation implements HeapSize, Comparable { * @param ts version timestamp * @param value column value * @return this - * @deprecated Since 1.0.0. Use {@link #addColumn(byte[], byte[], long, byte[])} - */ - @Deprecated - public Put add(byte [] family, byte [] qualifier, long ts, byte [] value) { - return addColumn(family, qualifier, ts, value); - } - - /** - * Add the specified column and value, with the specified timestamp as - * its version to this Put operation. - * @param family family name - * @param qualifier column qualifier - * @param ts version timestamp - * @param value column value - * @return this */ public Put addColumn(byte [] family, byte [] qualifier, long ts, byte [] value) { if (ts < 0) { @@ -211,7 +183,7 @@ public class Put extends Mutation implements HeapSize, Comparable { } /** - * See {@link #add(byte[], byte[], long, byte[])}. This version expects + * See {@link #addColumn(byte[], byte[], long, byte[])}. This version expects * that the underlying arrays won't change. It's intended * for usage internal HBase to and for advanced client applications. */ @@ -269,21 +241,6 @@ public class Put extends Mutation implements HeapSize, Comparable { * @param ts version timestamp * @param value column value * @return this - * @deprecated Since 1.0.0. Use {@link Put#addColumn(byte[], ByteBuffer, long, ByteBuffer)} - */ - @Deprecated - public Put add(byte[] family, ByteBuffer qualifier, long ts, ByteBuffer value) { - return addColumn(family, qualifier, ts, value); - } - - /** - * Add the specified column and value, with the specified timestamp as - * its version to this Put operation. - * @param family family name - * @param qualifier column qualifier - * @param ts version timestamp - * @param value column value - * @return this */ public Put addColumn(byte[] family, ByteBuffer qualifier, long ts, ByteBuffer value) { if (ts < 0) { @@ -297,7 +254,7 @@ public class Put extends Mutation implements HeapSize, Comparable { } /** - * See {@link #add(byte[], ByteBuffer, long, ByteBuffer)}. This version expects + * See {@link #addColumn(byte[], ByteBuffer, long, ByteBuffer)}. This version expects * that the underlying arrays won't change. It's intended * for usage internal HBase to and for advanced client applications. */ diff --git hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java index a20ca4f..b784f7a 100644 --- hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java +++ hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java @@ -1071,7 +1071,7 @@ public class TestAsyncProcess { throw new IllegalArgumentException("unknown " + regCnt); } - p.add(DUMMY_BYTES_1, DUMMY_BYTES_1, DUMMY_BYTES_1); + p.addColumn(DUMMY_BYTES_1, DUMMY_BYTES_1, DUMMY_BYTES_1); return p; } diff --git hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java index e941440..0a5a37f 100644 --- hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java +++ hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java @@ -703,7 +703,7 @@ public class TestClientNoCluster extends Configured implements Tool { for (int i = 0; i < namespaceSpan; i++) { byte [] b = format(rd.nextLong()); Put p = new Put(b); - p.add(HConstants.CATALOG_FAMILY, b, b); + p.addColumn(HConstants.CATALOG_FAMILY, b, b); mutator.mutate(p); if (i % printInterval == 0) { LOG.info("Put " + printInterval + "/" + stopWatch.elapsedMillis()); diff --git hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestOperation.java hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestOperation.java index 84069fb..fa9c4ad 100644 --- hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestOperation.java +++ hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestOperation.java @@ -337,7 +337,7 @@ public class TestOperation { // produce a Put operation Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); // get its JSON representation, and parse it json = put.toJSON(); parsedJSON = mapper.readValue(json, HashMap.class); diff --git hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestPutDotHas.java hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestPutDotHas.java index c269e62..b90374b 100644 --- hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestPutDotHas.java +++ hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestPutDotHas.java @@ -41,7 +41,7 @@ public class TestPutDotHas { @Before public void setUp() { - put.add(FAMILY_01, QUALIFIER_01, TS, VALUE_01); + put.addColumn(FAMILY_01, QUALIFIER_01, TS, VALUE_01); } @Test diff --git hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java index 07c1755..1dab633 100644 --- hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java +++ hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java @@ -89,7 +89,7 @@ public class IndexBuilder extends Configured implements Tool { // original: row 123 attribute:phone 555-1212 // index: row 555-1212 INDEX:ROW 123 Put put = new Put(value); - put.add(INDEX_COLUMN, INDEX_QUALIFIER, rowKey.get()); + put.addColumn(INDEX_COLUMN, INDEX_QUALIFIER, rowKey.get()); context.write(tableName, put); } } diff --git hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java index 011147a..18eb5a6 100644 --- hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java +++ hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java @@ -26,7 +26,6 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; -import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.Text; @@ -91,7 +90,7 @@ public class SampleUploader extends Configured implements Tool { // Create Put Put put = new Put(row); - put.add(family, qualifier, value); + put.addColumn(family, qualifier, value); // Uncomment below to disable WAL. This will improve performance but means // you will experience data loss in the case of a RegionServer crash. diff --git hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java index 511bf46..317081b 100644 --- hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java +++ hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java @@ -34,7 +34,6 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.testclassification.CoprocessorTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; @@ -224,8 +223,8 @@ public class TestBulkDeleteProtocol { List puts = new ArrayList(100); for (int j = 0; j < 100; j++) { Put put = new Put(Bytes.toBytes(j)); - put.add(FAMILY1, QUALIFIER1, "v1".getBytes()); - put.add(FAMILY2, QUALIFIER2, "v2".getBytes()); + put.addColumn(FAMILY1, QUALIFIER1, "v1".getBytes()); + put.addColumn(FAMILY2, QUALIFIER2, "v2".getBytes()); puts.add(put); } ht.put(puts); @@ -252,15 +251,15 @@ public class TestBulkDeleteProtocol { for (int j = 0; j < 100; j++) { Put put = new Put(Bytes.toBytes(j)); byte[] value = "v1".getBytes(); - put.add(FAMILY1, QUALIFIER1, 1234L, value); - put.add(FAMILY1, QUALIFIER2, 1234L, value); - put.add(FAMILY1, QUALIFIER3, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER1, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER2, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER3, 1234L, value); // Latest version values value = "v2".getBytes(); - put.add(FAMILY1, QUALIFIER1, value); - put.add(FAMILY1, QUALIFIER2, value); - put.add(FAMILY1, QUALIFIER3, value); - put.add(FAMILY1, null, value); + put.addColumn(FAMILY1, QUALIFIER1, value); + put.addColumn(FAMILY1, QUALIFIER2, value); + put.addColumn(FAMILY1, QUALIFIER3, value); + put.addColumn(FAMILY1, null, value); puts.add(put); } ht.put(puts); @@ -301,19 +300,19 @@ public class TestBulkDeleteProtocol { Put put = new Put(Bytes.toBytes(j)); // TS = 1000L byte[] value = "v1".getBytes(); - put.add(FAMILY1, QUALIFIER1, 1000L, value); - put.add(FAMILY1, QUALIFIER2, 1000L, value); - put.add(FAMILY1, QUALIFIER3, 1000L, value); + put.addColumn(FAMILY1, QUALIFIER1, 1000L, value); + put.addColumn(FAMILY1, QUALIFIER2, 1000L, value); + put.addColumn(FAMILY1, QUALIFIER3, 1000L, value); // TS = 1234L value = "v2".getBytes(); - put.add(FAMILY1, QUALIFIER1, 1234L, value); - put.add(FAMILY1, QUALIFIER2, 1234L, value); - put.add(FAMILY1, QUALIFIER3, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER1, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER2, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER3, 1234L, value); // Latest version values value = "v3".getBytes(); - put.add(FAMILY1, QUALIFIER1, value); - put.add(FAMILY1, QUALIFIER2, value); - put.add(FAMILY1, QUALIFIER3, value); + put.addColumn(FAMILY1, QUALIFIER1, value); + put.addColumn(FAMILY1, QUALIFIER2, value); + put.addColumn(FAMILY1, QUALIFIER3, value); puts.add(put); } ht.put(puts); @@ -348,24 +347,24 @@ public class TestBulkDeleteProtocol { Put put = new Put(Bytes.toBytes(j)); // TS = 1000L byte[] value = "v1".getBytes(); - put.add(FAMILY1, QUALIFIER1, 1000L, value); - put.add(FAMILY1, QUALIFIER2, 1000L, value); - put.add(FAMILY1, QUALIFIER3, 1000L, value); + put.addColumn(FAMILY1, QUALIFIER1, 1000L, value); + put.addColumn(FAMILY1, QUALIFIER2, 1000L, value); + put.addColumn(FAMILY1, QUALIFIER3, 1000L, value); // TS = 1234L value = "v2".getBytes(); - put.add(FAMILY1, QUALIFIER1, 1234L, value); - put.add(FAMILY1, QUALIFIER2, 1234L, value); - put.add(FAMILY1, QUALIFIER3, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER1, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER2, 1234L, value); + put.addColumn(FAMILY1, QUALIFIER3, 1234L, value); // TS = 2000L value = "v3".getBytes(); - put.add(FAMILY1, QUALIFIER1, 2000L, value); - put.add(FAMILY1, QUALIFIER2, 2000L, value); - put.add(FAMILY1, QUALIFIER3, 2000L, value); + put.addColumn(FAMILY1, QUALIFIER1, 2000L, value); + put.addColumn(FAMILY1, QUALIFIER2, 2000L, value); + put.addColumn(FAMILY1, QUALIFIER3, 2000L, value); // Latest version values value = "v4".getBytes(); - put.add(FAMILY1, QUALIFIER1, value); - put.add(FAMILY1, QUALIFIER2, value); - put.add(FAMILY1, QUALIFIER3, value); + put.addColumn(FAMILY1, QUALIFIER1, value); + put.addColumn(FAMILY1, QUALIFIER2, value); + put.addColumn(FAMILY1, QUALIFIER3, value); puts.add(put); } ht.put(puts); @@ -436,9 +435,9 @@ public class TestBulkDeleteProtocol { private Put createPut(byte[] rowkey, String value) throws IOException { Put put = new Put(rowkey); - put.add(FAMILY1, QUALIFIER1, value.getBytes()); - put.add(FAMILY1, QUALIFIER2, value.getBytes()); - put.add(FAMILY1, QUALIFIER3, value.getBytes()); + put.addColumn(FAMILY1, QUALIFIER1, value.getBytes()); + put.addColumn(FAMILY1, QUALIFIER2, value.getBytes()); + put.addColumn(FAMILY1, QUALIFIER3, value.getBytes()); return put; } } diff --git hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java index 481cb91..1776ced 100644 --- hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java +++ hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java @@ -21,7 +21,6 @@ package org.apache.hadoop.hbase.coprocessor.example; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.client.coprocessor.Batch; @@ -77,7 +76,7 @@ public class TestRowCountEndpoint { for (int i=0; i<5; i++) { byte[] iBytes = Bytes.toBytes(i); Put p = new Put(iBytes); - p.add(TEST_FAMILY, TEST_COLUMN, iBytes); + p.addColumn(TEST_FAMILY, TEST_COLUMN, iBytes); table.put(p); } diff --git hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java index db10c5a..e97d528 100644 --- hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java +++ hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.java @@ -28,7 +28,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Table; @@ -94,10 +93,10 @@ public class TestZooKeeperScanPolicyObserver { long ts = now - 2000; Put p = new Put(R); - p.add(F, Q, ts, Q); + p.addColumn(F, Q, ts, Q); t.put(p); p = new Put(R); - p.add(F, Q, ts+1, Q); + p.addColumn(F, Q, ts + 1, Q); t.put(p); // these two should be expired but for the override diff --git hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java index 4423650..437f200 100644 --- hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java +++ hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java @@ -496,7 +496,7 @@ public class IntegrationTestMTTR { @Override protected boolean doAction() throws Exception { Put p = new Put(Bytes.toBytes(RandomStringUtils.randomAlphanumeric(5))); - p.add(FAMILY, Bytes.toBytes("\0"), Bytes.toBytes(RandomStringUtils.randomAscii(5))); + p.addColumn(FAMILY, Bytes.toBytes("\0"), Bytes.toBytes(RandomStringUtils.randomAscii(5))); table.put(p); return true; } diff --git hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java index 3212ad4..2a146b3 100644 --- hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java +++ hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java @@ -211,13 +211,14 @@ public class IntegrationTestBigLinkedListWithVisibility extends IntegrationTestB for (int i = 0; i < current.length; i++) { for (int j = 0; j < DEFAULT_TABLES_COUNT; j++) { Put put = new Put(current[i]); - put.add(FAMILY_NAME, COLUMN_PREV, prev == null ? NO_KEY : prev[i]); + byte[] value = prev == null ? NO_KEY : prev[i]; + put.addColumn(FAMILY_NAME, COLUMN_PREV, value); if (count >= 0) { - put.add(FAMILY_NAME, COLUMN_COUNT, Bytes.toBytes(count + i)); + put.addColumn(FAMILY_NAME, COLUMN_COUNT, Bytes.toBytes(count + i)); } if (id != null) { - put.add(FAMILY_NAME, COLUMN_CLIENT, id); + put.addColumn(FAMILY_NAME, COLUMN_CLIENT, id); } visibilityExps = split[j * 2] + OR + split[(j * 2) + 1]; put.setCellVisibility(new CellVisibility(visibilityExps)); diff --git hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java index 623a370..e279dfb 100644 --- hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java +++ hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java @@ -242,12 +242,12 @@ public void cleanUpCluster() throws Exception { Bytes.putLong(row, 0, byteSwapped); Put p = new Put(row); - p.add(TEST_FAMILY, TEST_QUALIFIER, HConstants.EMPTY_BYTE_ARRAY); + p.addColumn(TEST_FAMILY, TEST_QUALIFIER, HConstants.EMPTY_BYTE_ARRAY); if (blockStart > 0) { for (int j = 0; j < numBackReferencesPerRow; j++) { long referredRow = blockStart - BLOCK_SIZE + rand.nextInt(BLOCK_SIZE); Bytes.putLong(row, 0, swapLong(referredRow)); - p.add(TEST_FAMILY, row, HConstants.EMPTY_BYTE_ARRAY); + p.addColumn(TEST_FAMILY, row, HConstants.EMPTY_BYTE_ARRAY); } refsWritten.increment(1); } diff --git hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java index b797740..52a705b 100644 --- hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java +++ hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java @@ -174,7 +174,7 @@ public class IntegrationTestWithCellVisibilityLoadAndVerify extends IntegrationT String exp = VISIBILITY_EXPS[expIdx]; byte[] row = Bytes.add(Bytes.toBytes(i), Bytes.toBytes(suffix), Bytes.toBytes(exp)); Put p = new Put(row); - p.add(TEST_FAMILY, TEST_QUALIFIER, HConstants.EMPTY_BYTE_ARRAY); + p.addColumn(TEST_FAMILY, TEST_QUALIFIER, HConstants.EMPTY_BYTE_ARRAY); p.setCellVisibility(new CellVisibility(exp)); getCounter(expIdx).increment(1); mutator.mutate(p); diff --git hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java index 3fa8a9c..f325aac 100644 --- hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java +++ hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java @@ -245,7 +245,7 @@ public class IntegrationTestSendTraceRequests extends AbstractHBaseTool { Put p = new Put(Bytes.toBytes(rk)); for (int y = 0; y < 10; y++) { random.nextBytes(value); - p.add(familyName, Bytes.toBytes(random.nextLong()), value); + p.addColumn(familyName, Bytes.toBytes(random.nextLong()), value); } ht.mutate(p); } diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/PerformanceEvaluation.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/PerformanceEvaluation.java index e91f873..8424bf9 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/PerformanceEvaluation.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/PerformanceEvaluation.java @@ -1131,7 +1131,7 @@ public class PerformanceEvaluation extends Configured implements Tool { value, tags); put.add(kv); } else { - put.add(FAMILY_NAME, QUALIFIER_NAME, value); + put.addColumn(FAMILY_NAME, QUALIFIER_NAME, value); } put.setDurability(writeToWAL ? Durability.SYNC_WAL : Durability.SKIP_WAL); mutator.mutate(put); @@ -1202,7 +1202,7 @@ public class PerformanceEvaluation extends Configured implements Tool { value, tags); put.add(kv); } else { - put.add(FAMILY_NAME, QUALIFIER_NAME, value); + put.addColumn(FAMILY_NAME, QUALIFIER_NAME, value); } put.setDurability(writeToWAL ? Durability.SYNC_WAL : Durability.SKIP_WAL); mutator.mutate(put); diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannerResource.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannerResource.java index 4f4f698..5114b11 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannerResource.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannerResource.java @@ -100,7 +100,7 @@ public class TestScannerResource { k[2] = b3; Put put = new Put(k); put.setDurability(Durability.SKIP_WAL); - put.add(famAndQf[0], famAndQf[1], k); + put.addColumn(famAndQf[0], famAndQf[1], k); puts.add(put); } } diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithFilters.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithFilters.java index ef54d9f..eed4f1a 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithFilters.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithFilters.java @@ -149,7 +149,7 @@ public class TestScannersWithFilters { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_ONE) { - p.add(FAMILIES[0], QUALIFIER, VALUES[0]); + p.addColumn(FAMILIES[0], QUALIFIER, VALUES[0]); } table.put(p); } @@ -157,7 +157,7 @@ public class TestScannersWithFilters { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_TWO) { - p.add(FAMILIES[1], QUALIFIER, VALUES[1]); + p.addColumn(FAMILIES[1], QUALIFIER, VALUES[1]); } table.put(p); } @@ -167,7 +167,7 @@ public class TestScannersWithFilters { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_ONE) { - p.add(FAMILIES[1], QUALIFIER, VALUES[0]); + p.addColumn(FAMILIES[1], QUALIFIER, VALUES[0]); } table.put(p); } @@ -175,7 +175,7 @@ public class TestScannersWithFilters { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_TWO) { - p.add(FAMILIES[0], QUALIFIER, VALUES[1]); + p.addColumn(FAMILIES[0], QUALIFIER, VALUES[1]); } table.put(p); } diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithLabels.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithLabels.java index 83c72e3..cb235aa 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithLabels.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestScannersWithLabels.java @@ -98,7 +98,7 @@ public class TestScannersWithLabels { for (int i = 0; i < 9; i++) { Put put = new Put(Bytes.toBytes("row" + i)); put.setDurability(Durability.SKIP_WAL); - put.add(famAndQf[0], famAndQf[1], k); + put.addColumn(famAndQf[0], famAndQf[1], k); put.setCellVisibility(new CellVisibility("(" + SECRET + "|" + CONFIDENTIAL + ")" + "&" + "!" + TOPSECRET)); puts.add(put); diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableResource.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableResource.java index b0b8fef..8fe5c98 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableResource.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableResource.java @@ -108,7 +108,7 @@ public class TestTableResource { k[2] = b3; Put put = new Put(k); put.setDurability(Durability.SKIP_WAL); - put.add(famAndQf[0], famAndQf[1], k); + put.addColumn(famAndQf[0], famAndQf[1], k); puts.add(put); } } diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteHTableRetries.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteHTableRetries.java index 5b18a6a..4a595f3 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteHTableRetries.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteHTableRetries.java @@ -151,7 +151,7 @@ public class TestRemoteHTableRetries { @Override public void run() throws Exception { Put put = new Put(ROW_1); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); remoteTable.checkAndPut(ROW_1, COLUMN_1, QUALIFIER_1, VALUE_1, put ); } }); @@ -164,7 +164,7 @@ public class TestRemoteHTableRetries { @Override public void run() throws Exception { Put put = new Put(ROW_1); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); Delete delete= new Delete(ROW_1); remoteTable.checkAndDelete(ROW_1, COLUMN_1, QUALIFIER_1, VALUE_1, delete ); } diff --git hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteTable.java hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteTable.java index 297162b..19d0587 100644 --- hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteTable.java +++ hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/client/TestRemoteTable.java @@ -100,12 +100,12 @@ public class TestRemoteTable { admin.createTable(htd); try (Table table = TEST_UTIL.getConnection().getTable(TABLE)) { Put put = new Put(ROW_1); - put.add(COLUMN_1, QUALIFIER_1, TS_2, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, TS_2, VALUE_1); table.put(put); put = new Put(ROW_2); - put.add(COLUMN_1, QUALIFIER_1, TS_1, VALUE_1); - put.add(COLUMN_1, QUALIFIER_1, TS_2, VALUE_2); - put.add(COLUMN_2, QUALIFIER_2, TS_2, VALUE_2); + put.addColumn(COLUMN_1, QUALIFIER_1, TS_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, TS_2, VALUE_2); + put.addColumn(COLUMN_2, QUALIFIER_2, TS_2, VALUE_2); table.put(put); } remoteTable = new RemoteHTable( @@ -282,7 +282,7 @@ public class TestRemoteTable { @Test public void testPut() throws IOException { Put put = new Put(ROW_3); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); remoteTable.put(put); Get get = new Get(ROW_3); @@ -296,13 +296,13 @@ public class TestRemoteTable { List puts = new ArrayList(); put = new Put(ROW_3); - put.add(COLUMN_2, QUALIFIER_2, VALUE_2); + put.addColumn(COLUMN_2, QUALIFIER_2, VALUE_2); puts.add(put); put = new Put(ROW_4); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_4); - put.add(COLUMN_2, QUALIFIER_2, VALUE_2); + put.addColumn(COLUMN_2, QUALIFIER_2, VALUE_2); puts.add(put); remoteTable.put(puts); @@ -327,8 +327,8 @@ public class TestRemoteTable { @Test public void testDelete() throws IOException { Put put = new Put(ROW_3); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); - put.add(COLUMN_2, QUALIFIER_2, VALUE_2); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_2, QUALIFIER_2, VALUE_2); remoteTable.put(put); Get get = new Get(ROW_3); @@ -390,16 +390,16 @@ public class TestRemoteTable { public void testScanner() throws IOException { List puts = new ArrayList(); Put put = new Put(ROW_1); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_2); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_3); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_4); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); remoteTable.put(puts); @@ -465,7 +465,7 @@ public class TestRemoteTable { assertFalse(remoteTable.exists(get)); Put put = new Put(ROW_1); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); remoteTable.put(put); assertTrue(remoteTable.checkAndPut(ROW_1, COLUMN_1, QUALIFIER_1, VALUE_1, @@ -481,16 +481,16 @@ public class TestRemoteTable { public void testIteratorScaner() throws IOException { List puts = new ArrayList(); Put put = new Put(ROW_1); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_2); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_3); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); put = new Put(ROW_4); - put.add(COLUMN_1, QUALIFIER_1, VALUE_1); + put.addColumn(COLUMN_1, QUALIFIER_1, VALUE_1); puts.add(put); remoteTable.put(puts); diff --git hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java index aaaabba..af64c66 100644 --- hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java +++ hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java @@ -144,7 +144,7 @@ public class QuotaUtil extends QuotaTableUtil { private static void addQuotas(final Connection connection, final byte[] rowKey, final byte[] qualifier, final Quotas data) throws IOException { Put put = new Put(rowKey); - put.add(QUOTA_FAMILY_INFO, qualifier, quotasToData(data)); + put.addColumn(QUOTA_FAMILY_INFO, qualifier, quotasToData(data)); doPut(connection, put); } diff --git hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java index 28cb9b9..6b8c1ee 100644 --- hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java +++ hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionMergeTransactionImpl.java @@ -410,10 +410,8 @@ public class RegionMergeTransactionImpl implements RegionMergeTransaction { // Put for parent Put putOfMerged = MetaTableAccessor.makePutFromRegionInfo(copyOfMerged, time); - putOfMerged.add(HConstants.CATALOG_FAMILY, HConstants.MERGEA_QUALIFIER, - regionA.toByteArray()); - putOfMerged.add(HConstants.CATALOG_FAMILY, HConstants.MERGEB_QUALIFIER, - regionB.toByteArray()); + putOfMerged.addColumn(HConstants.CATALOG_FAMILY, HConstants.MERGEA_QUALIFIER, regionA.toByteArray()); + putOfMerged.addColumn(HConstants.CATALOG_FAMILY, HConstants.MERGEB_QUALIFIER, regionB.toByteArray()); mutations.add(putOfMerged); // Deletes for merging regions Delete deleteA = MetaTableAccessor.makeDeleteFromRegionInfo(regionA, time); @@ -426,11 +424,11 @@ public class RegionMergeTransactionImpl implements RegionMergeTransaction { @VisibleForTesting Put addLocation(final Put p, final ServerName sn, long openSeqNum) { - p.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes - .toBytes(sn.getHostAndPort())); - p.add(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn - .getStartcode())); - p.add(HConstants.CATALOG_FAMILY, HConstants.SEQNUM_QUALIFIER, Bytes.toBytes(openSeqNum)); + p.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes + .toBytes(sn.getHostAndPort())); + p.addColumn(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn + .getStartcode())); + p.addColumn(HConstants.CATALOG_FAMILY, HConstants.SEQNUM_QUALIFIER, Bytes.toBytes(openSeqNum)); return p; } diff --git hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransactionImpl.java hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransactionImpl.java index fbfea8e..70d040e 100644 --- hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransactionImpl.java +++ hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransactionImpl.java @@ -329,11 +329,11 @@ public class SplitTransactionImpl implements SplitTransaction { @VisibleForTesting Put addLocation(final Put p, final ServerName sn, long openSeqNum) { - p.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes - .toBytes(sn.getHostAndPort())); - p.add(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn - .getStartcode())); - p.add(HConstants.CATALOG_FAMILY, HConstants.SEQNUM_QUALIFIER, Bytes.toBytes(openSeqNum)); + p.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes + .toBytes(sn.getHostAndPort())); + p.addColumn(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn + .getStartcode())); + p.addColumn(HConstants.CATALOG_FAMILY, HConstants.SEQNUM_QUALIFIER, Bytes.toBytes(openSeqNum)); return p; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java index 597573a..02a2a81 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java @@ -34,7 +34,6 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -348,9 +347,10 @@ public abstract class HBaseTestCase extends TestCase { byte[][] split = KeyValue.parseColumn(Bytes.toBytes(sb.toString())); if(split.length == 1) { - put.add(split[0], new byte[0], t); + byte[] qualifier = new byte[0]; + put.addColumn(split[0], qualifier, t); } else { - put.add(split[0], split[1], t); + put.addColumn(split[0], split[1], t); } put.setDurability(Durability.SKIP_WAL); updater.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index c6b5c37..befa06a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -1951,7 +1951,8 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { Put put = new Put(row); put.setDurability(writeToWAL ? Durability.USE_DEFAULT : Durability.SKIP_WAL); for (int i = 0; i < f.length; i++) { - put.add(f[i], null, value != null ? value : row); + byte[] value1 = value != null ? value : row; + put.addColumn(f[i], null, value1); } puts.add(put); } @@ -2038,7 +2039,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { k[2] = b3; Put put = new Put(k); put.setDurability(Durability.SKIP_WAL); - put.add(f, null, k); + put.addColumn(f, null, k); if (r.getWAL() == null) { put.setDurability(Durability.SKIP_WAL); } @@ -2068,7 +2069,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { for (int i = startRow; i < endRow; i++) { byte[] data = Bytes.toBytes(String.valueOf(i)); Put put = new Put(data); - put.add(f, null, data); + put.addColumn(f, null, data); t.put(put); } } @@ -3412,7 +3413,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { final byte[] value = Bytes.toBytes("value_for_row_" + iRow + "_cf_" + Bytes.toStringBinary(cf) + "_col_" + iCol + "_ts_" + ts + "_random_" + rand.nextLong()); - put.add(cf, qual, ts, value); + put.addColumn(cf, qual, ts, value); } else if (rand.nextDouble() < 0.8) { del.addColumn(cf, qual, ts); } else { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java index 569ef71..1c1c56c 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java @@ -1413,7 +1413,7 @@ public class PerformanceEvaluation extends Configured implements Tool { put.add(kv); updateValueSize(kv.getValueLength()); } else { - put.add(FAMILY_NAME, qualifier, value); + put.addColumn(FAMILY_NAME, qualifier, value); updateValueSize(value.length); } } @@ -1501,7 +1501,7 @@ public class PerformanceEvaluation extends Configured implements Tool { put.add(kv); updateValueSize(kv.getValueLength()); } else { - put.add(FAMILY_NAME, qualifier, value); + put.addColumn(FAMILY_NAME, qualifier, value); updateValueSize(value.length); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java index dbb6156..27b480d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java @@ -135,7 +135,7 @@ public class TestAcidGuarantees implements Tool { for (byte[] family : targetFamilies) { for (int i = 0; i < NUM_COLS_TO_CHECK; i++) { byte qualifier[] = Bytes.toBytes("col" + i); - p.add(family, qualifier, data); + p.addColumn(family, qualifier, data); } } table.put(p); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java index 3a9c565..dd09c37 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java @@ -98,7 +98,7 @@ public class TestHBaseTestingUtility { Table table2 = htu2.createTable(TABLE_NAME, FAM_NAME); Put put = new Put(ROW); - put.add(FAM_NAME, QUAL_NAME, VALUE); + put.addColumn(FAM_NAME, QUAL_NAME, VALUE); table1.put(put); Get get = new Get(ROW); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java index 278973e..155695a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java @@ -136,7 +136,7 @@ public class TestMultiVersions { desc.addFamily(hcd); this.admin.createTable(desc); Put put = new Put(row, timestamp1); - put.add(contents, contents, value1); + put.addColumn(contents, contents, value1); Table table = UTIL.getConnection().getTable(desc.getTableName()); table.put(put); // Shut down and restart the HBase cluster @@ -148,7 +148,7 @@ public class TestMultiVersions { table = UTIL.getConnection().getTable(desc.getTableName()); // Overwrite previous value put = new Put(row, timestamp2); - put.add(contents, contents, value2); + put.addColumn(contents, contents, value2); table.put(put); // Now verify that getRow(row, column, latest) works Get get = new Get(row); @@ -221,8 +221,7 @@ public class TestMultiVersions { for (int i = 0; i < startKeys.length; i++) { for (int j = 0; j < timestamp.length; j++) { Put put = new Put(rows[i], timestamp[j]); - put.add(HConstants.CATALOG_FAMILY, null, timestamp[j], - Bytes.toBytes(timestamp[j])); + put.addColumn(HConstants.CATALOG_FAMILY, null, timestamp[j], Bytes.toBytes(timestamp[j])); puts.add(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java index 050d403..c24d8a3 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java @@ -35,21 +35,16 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.MiscTests; import org.apache.hadoop.hbase.util.Bytes; -import org.apache.hadoop.hbase.util.FSUtils; -import org.apache.hadoop.hbase.zookeeper.ZKUtil; -import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.junit.AfterClass; import org.junit.Assert; import org.junit.Before; import org.junit.BeforeClass; -import org.junit.Ignore; import org.junit.Test; import org.junit.experimental.categories.Category; @@ -253,7 +248,7 @@ public class TestNamespace { //sanity check try to write and read from table Table table = TEST_UTIL.getConnection().getTable(desc.getTableName()); Put p = new Put(Bytes.toBytes("row1")); - p.add(Bytes.toBytes("my_cf"),Bytes.toBytes("my_col"),Bytes.toBytes("value1")); + p.addColumn(Bytes.toBytes("my_cf"), Bytes.toBytes("my_col"), Bytes.toBytes("value1")); table.put(p); //flush and read from disk to make sure directory changes are working admin.flush(desc.getTableName()); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java index ec3521c..58ac2cd 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java @@ -261,8 +261,7 @@ public class TestZooKeeper { Table table = TEST_UTIL.getConnection().getTable(desc.getTableName()); Put put = new Put(Bytes.toBytes("testrow")); - put.add(Bytes.toBytes("fam"), - Bytes.toBytes("col"), Bytes.toBytes("testdata")); + put.addColumn(Bytes.toBytes("fam"), Bytes.toBytes("col"), Bytes.toBytes("testdata")); LOG.info("Putting table " + tableName); table.put(put); table.close(); @@ -611,7 +610,7 @@ public class TestZooKeeper { int numberOfPuts; for (numberOfPuts = 0; numberOfPuts < 6; numberOfPuts++) { p = new Put(Bytes.toBytes(numberOfPuts)); - p.add(Bytes.toBytes("col"), Bytes.toBytes("ql"), Bytes.toBytes("value" + numberOfPuts)); + p.addColumn(Bytes.toBytes("col"), Bytes.toBytes("ql"), Bytes.toBytes("value" + numberOfPuts)); table.put(p); } m.getZooKeeper().close(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java hbase-server/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java index 5d0522b..ef3d7d2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java @@ -244,7 +244,7 @@ public class TimestampTestBase extends HBaseTestCase { throws IOException { Put put = new Put(ROW, ts); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_NAME, QUALIFIER_NAME, bytes); + put.addColumn(FAMILY_NAME, QUALIFIER_NAME, bytes); loader.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java index eba3c0b..a28112d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java @@ -403,7 +403,7 @@ public class TestZooKeeperTableArchiveClient { private void createHFileInRegion(Region region, byte[] columnFamily) throws IOException { // put one row in the region Put p = new Put(Bytes.toBytes("row")); - p.add(columnFamily, Bytes.toBytes("Qual"), Bytes.toBytes("v1")); + p.addColumn(columnFamily, Bytes.toBytes("Qual"), Bytes.toBytes("v1")); region.put(p); // flush the region to make a store file region.flush(true); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java index 33c151d..0ba51b0 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java @@ -241,7 +241,7 @@ public class TestAdmin1 { final TableName table = TableName.valueOf("testDisableAndEnableTable"); Table ht = TEST_UTIL.createTable(table, HConstants.CATALOG_FAMILY); Put put = new Put(row); - put.add(HConstants.CATALOG_FAMILY, qualifier, value); + put.addColumn(HConstants.CATALOG_FAMILY, qualifier, value); ht.put(put); Get get = new Get(row); get.addColumn(HConstants.CATALOG_FAMILY, qualifier); @@ -308,7 +308,7 @@ public class TestAdmin1 { Table ht1 = TEST_UTIL.createTable(table1, HConstants.CATALOG_FAMILY); Table ht2 = TEST_UTIL.createTable(table2, HConstants.CATALOG_FAMILY); Put put = new Put(row); - put.add(HConstants.CATALOG_FAMILY, qualifier, value); + put.addColumn(HConstants.CATALOG_FAMILY, qualifier, value); ht1.put(put); ht2.put(put); Get get = new Get(row); @@ -1187,13 +1187,13 @@ public class TestAdmin1 { List puts = new ArrayList(); byte[] qualifier = "c".getBytes(); Put put = new Put(new byte[]{(byte)'1'}); - put.add(cf, qualifier, "100".getBytes()); + put.addColumn(cf, qualifier, "100".getBytes()); puts.add(put); put = new Put(new byte[]{(byte)'6'}); - put.add(cf, qualifier, "100".getBytes()); + put.addColumn(cf, qualifier, "100".getBytes()); puts.add(put); put = new Put(new byte[]{(byte)'8'}); - put.add(cf, qualifier, "100".getBytes()); + put.addColumn(cf, qualifier, "100".getBytes()); puts.add(put); ht.put(puts); ht.close(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java index 33203fc..fe311a6 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java @@ -218,7 +218,7 @@ public class TestAdmin2 { byte[] value = Bytes.toBytes("somedata"); // This used to use an empty row... That must have been a bug Put put = new Put(value); - put.add(HConstants.CATALOG_FAMILY, HConstants.CATALOG_FAMILY, value); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.CATALOG_FAMILY, value); table.put(put); table.close(); } @@ -614,7 +614,7 @@ public class TestAdmin2 { HRegionServer regionServer = TEST_UTIL.getRSForFirstRegionInTable(tableName); for (int i = 1; i <= 256; i++) { // 256 writes should cause 8 log rolls Put put = new Put(Bytes.toBytes("row" + String.format("%1$04d", i))); - put.add(HConstants.CATALOG_FAMILY, null, value); + put.addColumn(HConstants.CATALOG_FAMILY, null, value); table.put(put); if (i % 32 == 0) { // After every 32 writes sleep to let the log roller run diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java index d175744..6dedee2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java @@ -183,10 +183,10 @@ public class TestBlockEvictionFromClient { // insert data. 2 Rows are added Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); assertTrue(Bytes.equals(table.get(new Get(ROW)).value(), data)); // data was in memstore so don't expect any changes @@ -214,7 +214,7 @@ public class TestBlockEvictionFromClient { byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); byte[] data2 = Bytes.add(data, data); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); Result r = table.get(new Get(ROW)); assertTrue(Bytes.equals(r.getValue(FAMILY, QUALIFIER), data)); @@ -332,16 +332,16 @@ public class TestBlockEvictionFromClient { BlockCache cache = cacheConf.getBlockCache(); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); region.flush(true); // flush the data @@ -389,16 +389,16 @@ public class TestBlockEvictionFromClient { regionName); BlockCache cache = setCacheProperties(region); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); for (int i = 1; i < 10; i++) { put = new Put(ROW); - put.add(FAMILY, Bytes.toBytes("testQualifier" + i), data2); + put.addColumn(FAMILY, Bytes.toBytes("testQualifier" + i), data2); table.put(put); if (i % 2 == 0) { region.flush(true); @@ -406,7 +406,7 @@ public class TestBlockEvictionFromClient { } byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); region.flush(true); // flush the data @@ -483,16 +483,16 @@ public class TestBlockEvictionFromClient { BlockCache cache = setCacheProperties(region); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); for (int i = 1; i < 10; i++) { put = new Put(ROW); - put.add(Bytes.toBytes("testFamily" + i), Bytes.toBytes("testQualifier" + i), data2); + put.addColumn(Bytes.toBytes("testFamily" + i), Bytes.toBytes("testQualifier" + i), data2); table.put(put); if (i % 2 == 0) { region.flush(true); @@ -501,7 +501,7 @@ public class TestBlockEvictionFromClient { region.flush(true); byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); region.flush(true); // flush the data @@ -577,16 +577,16 @@ public class TestBlockEvictionFromClient { BlockCache cache = cacheConf.getBlockCache(); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); region.flush(true); // flush the data @@ -656,16 +656,16 @@ public class TestBlockEvictionFromClient { BlockCache cache = setCacheProperties(region); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); region.flush(true); for (int i = 1; i < 10; i++) { put = new Put(ROW); - put.add(Bytes.toBytes("testFamily" + i), Bytes.toBytes("testQualifier" + i), data2); + put.addColumn(Bytes.toBytes("testFamily" + i), Bytes.toBytes("testQualifier" + i), data2); table.put(put); if (i % 2 == 0) { region.flush(true); @@ -674,7 +674,7 @@ public class TestBlockEvictionFromClient { region.flush(true); byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); region.flush(true); // flush the data @@ -827,10 +827,10 @@ public class TestBlockEvictionFromClient { // insert data. 2 Rows are added Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); assertTrue(Bytes.equals(table.get(new Get(ROW)).value(), data)); // Should create one Hfile with 2 blocks @@ -842,7 +842,7 @@ public class TestBlockEvictionFromClient { byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); byte[] data2 = Bytes.add(data, data); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); // flush, one new block System.out.println("Flushing cache"); @@ -1042,14 +1042,14 @@ public class TestBlockEvictionFromClient { private void insertData(HTable table) throws IOException { Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); put = new Put(ROW1); - put.add(FAMILY, QUALIFIER, data); + put.addColumn(FAMILY, QUALIFIER, data); table.put(put); byte[] QUALIFIER2 = Bytes.add(QUALIFIER, QUALIFIER); put = new Put(ROW); - put.add(FAMILY, QUALIFIER2, data2); + put.addColumn(FAMILY, QUALIFIER2, data2); table.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestCheckAndMutate.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestCheckAndMutate.java index b68381f..082de09 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestCheckAndMutate.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestCheckAndMutate.java @@ -62,9 +62,9 @@ public class TestCheckAndMutate { try { // put one row Put put = new Put(rowKey); - put.add(family, Bytes.toBytes("A"), Bytes.toBytes("a")); - put.add(family, Bytes.toBytes("B"), Bytes.toBytes("b")); - put.add(family, Bytes.toBytes("C"), Bytes.toBytes("c")); + put.addColumn(family, Bytes.toBytes("A"), Bytes.toBytes("a")); + put.addColumn(family, Bytes.toBytes("B"), Bytes.toBytes("b")); + put.addColumn(family, Bytes.toBytes("C"), Bytes.toBytes("c")); table.put(put); // get row back and assert the values Get get = new Get(rowKey); @@ -102,7 +102,8 @@ public class TestCheckAndMutate { //Test that we get a region level exception try { Put p = new Put(rowKey); - p.add(new byte[]{'b', 'o', 'g', 'u', 's'}, new byte[]{'A'}, new byte[0]); + byte[] value = new byte[0]; + p.addColumn(new byte[]{'b', 'o', 'g', 'u', 's'}, new byte[]{'A'}, value); rm = new RowMutations(rowKey); rm.add(p); table.checkAndMutate(rowKey, family, Bytes.toBytes("A"), CompareFilter.CompareOp.EQUAL, diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java index 072098e..a4603b7 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java @@ -85,7 +85,7 @@ public class TestClientOperationInterrupt { Table ht = util.createTable(tableName, new byte[][]{dummy, test}); Put p = new Put(row1); - p.add(dummy, dummy, dummy); + p.addColumn(dummy, dummy, dummy); ht.put(p); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java index 8af1d9f..515e763 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java @@ -126,7 +126,7 @@ public class TestClientScannerRPCTimeout { private void putToTable(Table ht, byte[] rowkey) throws IOException { Put put = new Put(rowkey); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java index 3ce7afb..5ceef01 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java @@ -35,7 +35,6 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; @@ -126,7 +125,7 @@ public class TestFastFail { byte[] rowKey = longToByteArrayKey(i); Put put = new Put(rowKey); byte[] value = rowKey; // value is the same as the row key - put.add(FAMILY, QUALIFIER, value); + put.addColumn(FAMILY, QUALIFIER, value); puts.add(put); } try (Table table = connection.getTable(TableName.valueOf(tableName))) { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java index e4e070b..79812cf 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java @@ -193,13 +193,13 @@ public class TestFromClientSide { long ts = System.currentTimeMillis(); Put p = new Put(T1, ts); - p.add(FAMILY, C0, T1); + p.addColumn(FAMILY, C0, T1); h.put(p); p = new Put(T1, ts+2); - p.add(FAMILY, C0, T2); + p.addColumn(FAMILY, C0, T2); h.put(p); p = new Put(T1, ts+4); - p.add(FAMILY, C0, T3); + p.addColumn(FAMILY, C0, T3); h.put(p); Delete d = new Delete(T1, ts+3); @@ -254,7 +254,7 @@ public class TestFromClientSide { // future timestamp long ts = System.currentTimeMillis() * 2; Put put = new Put(ROW, ts); - put.add(FAMILY, COLUMN, VALUE); + put.addColumn(FAMILY, COLUMN, VALUE); table.put(put); Get get = new Get(ROW); @@ -283,7 +283,7 @@ public class TestFromClientSide { }); put = new Put(ROW, ts); - put.add(FAMILY, COLUMN, VALUE); + put.addColumn(FAMILY, COLUMN, VALUE); table.put(put); get = new Get(ROW); @@ -418,15 +418,15 @@ public class TestFromClientSide { value)); Put put = new Put(Bytes.toBytes(row)); put.setDurability(Durability.SKIP_WAL); - put.add(Bytes.toBytes("trans-blob"), null, Bytes - .toBytes("value for blob")); - put.add(Bytes.toBytes("trans-type"), null, Bytes.toBytes("statement")); - put.add(Bytes.toBytes("trans-date"), null, Bytes - .toBytes("20090921010101999")); - put.add(Bytes.toBytes("trans-tags"), Bytes.toBytes("qual2"), Bytes - .toBytes(value)); - put.add(Bytes.toBytes("trans-group"), null, Bytes - .toBytes("adhocTransactionGroupId")); + put.addColumn(Bytes.toBytes("trans-blob"), null, Bytes + .toBytes("value for blob")); + put.addColumn(Bytes.toBytes("trans-type"), null, Bytes.toBytes("statement")); + put.addColumn(Bytes.toBytes("trans-date"), null, Bytes + .toBytes("20090921010101999")); + put.addColumn(Bytes.toBytes("trans-tags"), Bytes.toBytes("qual2"), Bytes + .toBytes(value)); + put.addColumn(Bytes.toBytes("trans-group"), null, Bytes + .toBytes("adhocTransactionGroupId")); ht.put(put); } } @@ -590,7 +590,7 @@ public class TestFromClientSide { TableName TABLE = TableName.valueOf("testSuperSimple"); Table ht = TEST_UTIL.createTable(TABLE, FAMILY); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); Scan scan = new Scan(); scan.addColumn(FAMILY, TABLE.toBytes()); @@ -608,7 +608,7 @@ public class TestFromClientSide { Table ht = TEST_UTIL.createTable(TABLE, FAMILY); byte[] value = new byte[4 * 1024 * 1024]; Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, value); + put.addColumn(FAMILY, QUALIFIER, value); ht.put(put); try { TEST_UTIL.getConfiguration().setInt( @@ -618,7 +618,7 @@ public class TestFromClientSide { ConnectionFactory.createConnection(TEST_UTIL.getConfiguration())) { try (Table t = connection.getTable(TableName.valueOf(FAMILY))) { put = new Put(ROW); - put.add(FAMILY, QUALIFIER, value); + put.addColumn(FAMILY, QUALIFIER, value); t.put(put); } } @@ -642,7 +642,7 @@ public class TestFromClientSide { for(int i=0;i<10;i++) { Put put = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY, QUALIFIERS[i], VALUE); + put.addColumn(FAMILY, QUALIFIERS[i], VALUE); ht.put(put); } Scan scan = new Scan(); @@ -675,7 +675,7 @@ public class TestFromClientSide { for(int i = 0; i < 10; i ++) { Put put = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY, QUALIFIER, values[i]); + put.addColumn(FAMILY, QUALIFIER, values[i]); ht.put(put); } Scan scan = new Scan(); @@ -709,7 +709,7 @@ public class TestFromClientSide { for(int i=0;i<10;i++) { Put put = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY, QUALIFIERS[i], VALUE); + put.addColumn(FAMILY, QUALIFIERS[i], VALUE); ht.put(put); } Scan scan = new Scan(); @@ -778,7 +778,7 @@ public class TestFromClientSide { // Insert a row Put put = new Put(ROWS[2]); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); // Try to get empty rows around it @@ -863,7 +863,7 @@ public class TestFromClientSide { //////////////////////////////////////////////////////////////////////////// put = new Put(ROWS[0]); - put.add(FAMILIES[4], QUALIFIERS[0], VALUES[0]); + put.addColumn(FAMILIES[4], QUALIFIERS[0], VALUES[0]); ht.put(put); // Get the single column @@ -896,13 +896,13 @@ public class TestFromClientSide { // Insert multiple columns to two other families put = new Put(ROWS[0]); - put.add(FAMILIES[2], QUALIFIERS[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIERS[4], VALUES[4]); - put.add(FAMILIES[4], QUALIFIERS[4], VALUES[4]); - put.add(FAMILIES[6], QUALIFIERS[6], VALUES[6]); - put.add(FAMILIES[6], QUALIFIERS[7], VALUES[7]); - put.add(FAMILIES[7], QUALIFIERS[7], VALUES[7]); - put.add(FAMILIES[9], QUALIFIERS[0], VALUES[0]); + put.addColumn(FAMILIES[2], QUALIFIERS[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIERS[4], VALUES[4]); + put.addColumn(FAMILIES[4], QUALIFIERS[4], VALUES[4]); + put.addColumn(FAMILIES[6], QUALIFIERS[6], VALUES[6]); + put.addColumn(FAMILIES[6], QUALIFIERS[7], VALUES[7]); + put.addColumn(FAMILIES[7], QUALIFIERS[7], VALUES[7]); + put.addColumn(FAMILIES[9], QUALIFIERS[0], VALUES[0]); ht.put(put); // Get multiple columns across multiple families and get empties around it @@ -923,10 +923,10 @@ public class TestFromClientSide { // Insert more data to memstore put = new Put(ROWS[0]); - put.add(FAMILIES[6], QUALIFIERS[5], VALUES[5]); - put.add(FAMILIES[6], QUALIFIERS[8], VALUES[8]); - put.add(FAMILIES[6], QUALIFIERS[9], VALUES[9]); - put.add(FAMILIES[4], QUALIFIERS[3], VALUES[3]); + put.addColumn(FAMILIES[6], QUALIFIERS[5], VALUES[5]); + put.addColumn(FAMILIES[6], QUALIFIERS[8], VALUES[8]); + put.addColumn(FAMILIES[6], QUALIFIERS[9], VALUES[9]); + put.addColumn(FAMILIES[4], QUALIFIERS[3], VALUES[3]); ht.put(put); //////////////////////////////////////////////////////////////////////////// @@ -1164,7 +1164,7 @@ public class TestFromClientSide { // Null row (should NOT work) try { Put put = new Put((byte[])null); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); fail("Inserting a null row worked, should throw exception"); } catch(Exception e) {} @@ -1172,7 +1172,7 @@ public class TestFromClientSide { // Null qualifier (should work) { Put put = new Put(ROW); - put.add(FAMILY, null, VALUE); + put.addColumn(FAMILY, null, VALUE); ht.put(put); getTestNull(ht, ROW, FAMILY, VALUE); @@ -1195,7 +1195,7 @@ public class TestFromClientSide { // Empty qualifier, byte[0] instead of null (should work) try { Put put = new Put(ROW); - put.add(FAMILY, HConstants.EMPTY_BYTE_ARRAY, VALUE); + put.addColumn(FAMILY, HConstants.EMPTY_BYTE_ARRAY, VALUE); ht.put(put); getTestNull(ht, ROW, FAMILY, VALUE); @@ -1225,7 +1225,7 @@ public class TestFromClientSide { // Null value try { Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, null); + put.addColumn(FAMILY, QUALIFIER, null); ht.put(put); Get get = new Get(ROW); @@ -1262,10 +1262,10 @@ public class TestFromClientSide { // Insert 4 versions of same column Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, STAMPS[1], VALUES[1]); - put.add(FAMILY, QUALIFIER, STAMPS[2], VALUES[2]); - put.add(FAMILY, QUALIFIER, STAMPS[4], VALUES[4]); - put.add(FAMILY, QUALIFIER, STAMPS[5], VALUES[5]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[1], VALUES[1]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[2], VALUES[2]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[4], VALUES[4]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[5], VALUES[5]); ht.put(put); // Verify we can get each one properly @@ -1351,10 +1351,10 @@ public class TestFromClientSide { // Insert 4 more versions of same column and a dupe put = new Put(ROW); - put.add(FAMILY, QUALIFIER, STAMPS[3], VALUES[3]); - put.add(FAMILY, QUALIFIER, STAMPS[6], VALUES[6]); - put.add(FAMILY, QUALIFIER, STAMPS[7], VALUES[7]); - put.add(FAMILY, QUALIFIER, STAMPS[8], VALUES[8]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[3], VALUES[3]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[6], VALUES[6]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[7], VALUES[7]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[8], VALUES[8]); ht.put(put); // Ensure maxVersions in query is respected @@ -1414,10 +1414,10 @@ public class TestFromClientSide { // Insert 4 more versions of same column and a dupe put = new Put(ROW); - put.add(FAMILY, QUALIFIER, STAMPS[9], VALUES[9]); - put.add(FAMILY, QUALIFIER, STAMPS[11], VALUES[11]); - put.add(FAMILY, QUALIFIER, STAMPS[13], VALUES[13]); - put.add(FAMILY, QUALIFIER, STAMPS[15], VALUES[15]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[9], VALUES[9]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[11], VALUES[11]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[13], VALUES[13]); + put.addColumn(FAMILY, QUALIFIER, STAMPS[15], VALUES[15]); ht.put(put); get = new Get(ROW); @@ -1476,19 +1476,19 @@ public class TestFromClientSide { // Insert limit + 1 on each family Put put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, STAMPS[0], VALUES[0]); - put.add(FAMILIES[0], QUALIFIER, STAMPS[1], VALUES[1]); - put.add(FAMILIES[1], QUALIFIER, STAMPS[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, STAMPS[1], VALUES[1]); - put.add(FAMILIES[1], QUALIFIER, STAMPS[2], VALUES[2]); - put.add(FAMILIES[1], QUALIFIER, STAMPS[3], VALUES[3]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[0], VALUES[0]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[3], VALUES[3]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[4], VALUES[4]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[5], VALUES[5]); - put.add(FAMILIES[2], QUALIFIER, STAMPS[6], VALUES[6]); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[0], VALUES[0]); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[1], VALUES[1]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[1], VALUES[1]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[2], VALUES[2]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[3], VALUES[3]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[0], VALUES[0]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[3], VALUES[3]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[4], VALUES[4]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[5], VALUES[5]); + put.addColumn(FAMILIES[2], QUALIFIER, STAMPS[6], VALUES[6]); ht.put(put); // Verify we only get the right number out of each @@ -1673,7 +1673,7 @@ public class TestFromClientSide { Put put = new Put(ROW); for (int q = 0; q < 1; q++) for (int t = 0; t < 5; t++) - put.add(FAMILY, QUALIFIERS[q], ts[t], VALUES[t]); + put.addColumn(FAMILY, QUALIFIERS[q], ts[t], VALUES[t]); ht.put(put); admin.flush(TABLE); @@ -1717,7 +1717,7 @@ public class TestFromClientSide { put = new Put(ROW); for (int q = 0; q < 5; q++) for (int t = 0; t < 5; t++) - put.add(FAMILY, QUALIFIERS[q], ts[t], VALUES[t]); + put.addColumn(FAMILY, QUALIFIERS[q], ts[t], VALUES[t]); ht.put(put); admin.flush(TABLE); @@ -1726,7 +1726,7 @@ public class TestFromClientSide { put = new Put(ROW2); for (int q = 0; q < 5; q++) for (int t = 0; t < 5; t++) - put.add(FAMILY, QUALIFIERS[q], ts[t], VALUES[t]); + put.addColumn(FAMILY, QUALIFIERS[q], ts[t], VALUES[t]); ht.put(put); admin.flush(TABLE); @@ -1825,8 +1825,8 @@ public class TestFromClientSide { Table ht = TEST_UTIL.createTable(TABLE, FAMILIES, 3); Put put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[0], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[1], VALUES[1]); ht.put(put); Delete delete = new Delete(ROW); @@ -1853,12 +1853,12 @@ public class TestFromClientSide { // Test delete latest version put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); - put.add(FAMILIES[0], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[0], QUALIFIER, ts[3], VALUES[3]); - put.add(FAMILIES[0], null, ts[4], VALUES[4]); - put.add(FAMILIES[0], null, ts[2], VALUES[2]); - put.add(FAMILIES[0], null, ts[3], VALUES[3]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[0], null, ts[4], VALUES[4]); + put.addColumn(FAMILIES[0], null, ts[2], VALUES[2]); + put.addColumn(FAMILIES[0], null, ts[3], VALUES[3]); ht.put(put); delete = new Delete(ROW); @@ -1897,8 +1897,8 @@ public class TestFromClientSide { // But alas, this is not to be. We can't put them back in either case. put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); // 1000 - put.add(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); // 5000 + put.addColumn(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); ht.put(put); @@ -1928,24 +1928,24 @@ public class TestFromClientSide { // Test deleting an entire family from one row but not the other various ways put = new Put(ROWS[0]); - put.add(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); ht.put(put); put = new Put(ROWS[1]); - put.add(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); ht.put(put); put = new Put(ROWS[2]); - put.add(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); ht.put(put); // Assert that above went in. @@ -2040,12 +2040,12 @@ public class TestFromClientSide { ht.delete(delete); put = new Put(ROWS[3]); - put.add(FAMILIES[2], QUALIFIER, VALUES[0]); + put.addColumn(FAMILIES[2], QUALIFIER, VALUES[0]); ht.put(put); put = new Put(ROWS[4]); - put.add(FAMILIES[1], QUALIFIER, VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, VALUES[2]); + put.addColumn(FAMILIES[1], QUALIFIER, VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, VALUES[2]); ht.put(put); get = new Get(ROWS[3]); @@ -2088,7 +2088,7 @@ public class TestFromClientSide { byte [] bytes = Bytes.toBytes(i); put = new Put(bytes); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILIES[0], QUALIFIER, bytes); + put.addColumn(FAMILIES[0], QUALIFIER, bytes); ht.put(put); } for (int i = 0; i < 10; i++) { @@ -2197,7 +2197,7 @@ public class TestFromClientSide { Put put = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); for(int j=0;j puts = new ArrayList(); for (int i=0;i !=100; i++){ Put put = new Put(Bytes.toBytes(i)); - put.add(FAMILY, FAMILY, Bytes.toBytes(i)); + put.addColumn(FAMILY, FAMILY, Bytes.toBytes(i)); puts.add(put); } foo.put(puts); @@ -5369,11 +5369,11 @@ public class TestFromClientSide { public void testScan_NullQualifier() throws IOException { Table table = TEST_UTIL.createTable(TableName.valueOf("testScan_NullQualifier"), FAMILY); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); put = new Put(ROW); - put.add(FAMILY, null, VALUE); + put.addColumn(FAMILY, null, VALUE); table.put(put); LOG.info("Row put"); @@ -5400,7 +5400,7 @@ public class TestFromClientSide { try { Put put = new Put(ROW, -1); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); fail("Negative timestamps should not have been allowed"); } catch (IllegalArgumentException ex) { @@ -5409,7 +5409,8 @@ public class TestFromClientSide { try { Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, -1, VALUE); + long ts = -1; + put.addColumn(FAMILY, QUALIFIER, ts, VALUE); table.put(put); fail("Negative timestamps should not have been allowed"); } catch (IllegalArgumentException ex) { @@ -5606,18 +5607,18 @@ public class TestFromClientSide { // put the same row 4 times, with different values Put p = new Put(row); - p.add(FAMILY, QUALIFIER, 10, VALUE); + p.addColumn(FAMILY, QUALIFIER, (long) 10, VALUE); table.put(p); p = new Put(row); - p.add(FAMILY, QUALIFIER, 11, ArrayUtils.add(VALUE, (byte) 2)); + p.addColumn(FAMILY, QUALIFIER, (long) 11, ArrayUtils.add(VALUE, (byte) 2)); table.put(p); p = new Put(row); - p.add(FAMILY, QUALIFIER, 12, ArrayUtils.add(VALUE, (byte) 3)); + p.addColumn(FAMILY, QUALIFIER, (long) 12, ArrayUtils.add(VALUE, (byte) 3)); table.put(p); p = new Put(row); - p.add(FAMILY, QUALIFIER, 13, ArrayUtils.add(VALUE, (byte) 4)); + p.addColumn(FAMILY, QUALIFIER, (long) 13, ArrayUtils.add(VALUE, (byte) 4)); table.put(p); int versions = 4; @@ -5678,7 +5679,7 @@ public class TestFromClientSide { int insertNum = 10; for (int i = 0; i < 10; i++) { Put put = new Put(Bytes.toBytes("row" + String.format("%03d", i))); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); } @@ -5710,34 +5711,34 @@ public class TestFromClientSide { TableName TABLE = TableName.valueOf("testSuperSimpleWithReverseScan"); Table ht = TEST_UTIL.createTable(TABLE, FAMILY); Put put = new Put(Bytes.toBytes("0-b11111-0000000000000000000")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b11111-0000000000000000002")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b11111-0000000000000000004")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b11111-0000000000000000006")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b11111-0000000000000000008")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b22222-0000000000000000001")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b22222-0000000000000000003")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b22222-0000000000000000005")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b22222-0000000000000000007")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); put = new Put(Bytes.toBytes("0-b22222-0000000000000000009")); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); Scan scan = new Scan(Bytes.toBytes("0-b11111-9223372036854775807"), Bytes.toBytes("0-b11111-0000000000000000000")); @@ -5767,7 +5768,7 @@ public class TestFromClientSide { Bytes.toBytes("col9--") }; for (int i = 0; i < 10; i++) { Put put = new Put(ROWS[i]); - put.add(FAMILY, QUALIFIERS[i], VALUE); + put.addColumn(FAMILY, QUALIFIERS[i], VALUE); ht.put(put); } Scan scan = new Scan(); @@ -5809,7 +5810,7 @@ public class TestFromClientSide { Bytes.toBytes("col9--") }; for (int i = 0; i < 10; i++) { Put put = new Put(ROWS[i]); - put.add(FAMILY, QUALIFIERS[i], VALUE); + put.addColumn(FAMILY, QUALIFIERS[i], VALUE); ht.put(put); } Scan scan = new Scan(); @@ -5870,7 +5871,7 @@ public class TestFromClientSide { // Insert a row Put put = new Put(ROWS[2]); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); ht.put(put); // Make sure we can scan the row @@ -5904,7 +5905,7 @@ public class TestFromClientSide { Table ht = TEST_UTIL.createTable(TABLE, FAMILY); // Null qualifier (should work) Put put = new Put(ROW); - put.add(FAMILY, null, VALUE); + put.addColumn(FAMILY, null, VALUE); ht.put(put); scanTestNull(ht, ROW, FAMILY, VALUE, true); Delete delete = new Delete(ROW); @@ -5915,7 +5916,7 @@ public class TestFromClientSide { ht = TEST_UTIL.createTable(TableName.valueOf(TABLE2), FAMILY); // Empty qualifier, byte[0] instead of null (should work) put = new Put(ROW); - put.add(FAMILY, HConstants.EMPTY_BYTE_ARRAY, VALUE); + put.addColumn(FAMILY, HConstants.EMPTY_BYTE_ARRAY, VALUE); ht.put(put); scanTestNull(ht, ROW, FAMILY, VALUE, true); TEST_UTIL.flush(); @@ -5925,7 +5926,7 @@ public class TestFromClientSide { ht.delete(delete); // Null value put = new Put(ROW); - put.add(FAMILY, QUALIFIER, null); + put.addColumn(FAMILY, QUALIFIER, null); ht.put(put); Scan scan = new Scan(); scan.setReversed(true); @@ -5945,8 +5946,8 @@ public class TestFromClientSide { Table ht = TEST_UTIL.createTable(TABLE, FAMILIES, 3); Put put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[0], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[1], VALUES[1]); ht.put(put); Delete delete = new Delete(ROW); @@ -5963,12 +5964,12 @@ public class TestFromClientSide { // Test delete latest version put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); - put.add(FAMILIES[0], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[0], QUALIFIER, ts[3], VALUES[3]); - put.add(FAMILIES[0], null, ts[4], VALUES[4]); - put.add(FAMILIES[0], null, ts[2], VALUES[2]); - put.add(FAMILIES[0], null, ts[3], VALUES[3]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[0], null, ts[4], VALUES[4]); + put.addColumn(FAMILIES[0], null, ts[2], VALUES[2]); + put.addColumn(FAMILIES[0], null, ts[3], VALUES[3]); ht.put(put); delete = new Delete(ROW); @@ -5997,8 +5998,8 @@ public class TestFromClientSide { // But alas, this is not to be. We can't put them back in either case. put = new Put(ROW); - put.add(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); // 1000 - put.add(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); // 5000 + put.addColumn(FAMILIES[0], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[0], QUALIFIER, ts[4], VALUES[4]); ht.put(put); // The Scanner returns the previous values, the expected-naive-unexpected @@ -6016,24 +6017,24 @@ public class TestFromClientSide { // ways put = new Put(ROWS[0]); - put.add(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); ht.put(put); put = new Put(ROWS[1]); - put.add(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); ht.put(put); put = new Put(ROWS[2]); - put.add(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); - put.add(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); - put.add(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[0], VALUES[0]); + put.addColumn(FAMILIES[1], QUALIFIER, ts[1], VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[2], VALUES[2]); + put.addColumn(FAMILIES[2], QUALIFIER, ts[3], VALUES[3]); ht.put(put); delete = new Delete(ROWS[0]); @@ -6087,12 +6088,12 @@ public class TestFromClientSide { ht.delete(delete); put = new Put(ROWS[3]); - put.add(FAMILIES[2], QUALIFIER, VALUES[0]); + put.addColumn(FAMILIES[2], QUALIFIER, VALUES[0]); ht.put(put); put = new Put(ROWS[4]); - put.add(FAMILIES[1], QUALIFIER, VALUES[1]); - put.add(FAMILIES[2], QUALIFIER, VALUES[2]); + put.addColumn(FAMILIES[1], QUALIFIER, VALUES[1]); + put.addColumn(FAMILIES[2], QUALIFIER, VALUES[2]); ht.put(put); scan = new Scan(ROWS[4]); @@ -6142,7 +6143,7 @@ public class TestFromClientSide { int insertNum = splitRows.length; for (int i = 0; i < insertNum; i++) { Put put = new Put(splitRows[i]); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); } @@ -6194,14 +6195,14 @@ public class TestFromClientSide { } for (byte[] splitRow : splitRows) { Put put = new Put(splitRow); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); byte[] nextRow = Bytes.copy(splitRow); nextRow[nextRow.length - 1]++; put = new Put(nextRow); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index 22309ef..ece98c1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -109,7 +109,7 @@ public class TestFromClientSide3 { for (int i = 0; i < nPuts; i++) { byte[] qualifier = Bytes.toBytes(random.nextInt()); byte[] value = Bytes.toBytes(random.nextInt()); - put.add(family, qualifier, value); + put.addColumn(family, qualifier, value); } table.put(put); } @@ -274,7 +274,7 @@ public class TestFromClientSide3 { actions.add(put1); Put put2 = new Put(ANOTHERROW); - put2.add(FAMILY, QUALIFIER, VALUE); + put2.addColumn(FAMILY, QUALIFIER, VALUE); actions.add(put2); table.batch(actions, results); @@ -294,7 +294,7 @@ public class TestFromClientSide3 { new byte[][] { FAMILY }); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); Get get = new Get(ROW); @@ -312,7 +312,7 @@ public class TestFromClientSide3 { "testHTableExistsMethodSingleRegionMultipleGets"), new byte[][] { FAMILY }); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); List gets = new ArrayList(); @@ -406,7 +406,7 @@ public class TestFromClientSide3 { TableName.valueOf("testHTableExistsMethodMultipleRegionsMultipleGets"), new byte[][] { FAMILY }, 1, new byte[] { 0x00 }, new byte[] { (byte) 0xff }, 255); Put put = new Put(ROW); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put (put); List gets = new ArrayList(); @@ -424,7 +424,7 @@ public class TestFromClientSide3 { // Test with the first region. put = new Put(new byte[] { 0x00 }); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); gets = new ArrayList(); @@ -436,7 +436,7 @@ public class TestFromClientSide3 { // Test with the last region put = new Put(new byte[] { (byte) 0xff, (byte) 0xff }); - put.add(FAMILY, QUALIFIER, VALUE); + put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); gets = new ArrayList(); @@ -459,7 +459,7 @@ public class TestFromClientSide3 { Table table = TEST_UTIL.getConnection().getTable(desc.getTableName()); Put put = new Put(ROW_BYTES); - put.add(FAMILY, COL_QUAL, VAL_BYTES); + put.addColumn(FAMILY, COL_QUAL, VAL_BYTES); table.put(put); //Try getting the row with an empty row key diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSideNoCodec.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSideNoCodec.java index dbb1cd1..550a4c9 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSideNoCodec.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSideNoCodec.java @@ -69,7 +69,9 @@ public class TestFromClientSideNoCodec { // Check put and get. final byte [] row = Bytes.toBytes("row"); Put p = new Put(row); - for (byte [] f: fs) p.add(f, f, f); + for (byte [] f: fs) { + p.addColumn(f, f, f); + } ht.put(p); Result r = ht.get(new Get(row)); int i = 0; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java index 041b19f..7e5a4bf 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java @@ -248,7 +248,7 @@ public class TestHCM { } Put p1 = new Put(rk); - p1.add(cf, "qual".getBytes(), "val".getBytes()); + p1.addColumn(cf, "qual".getBytes(), "val".getBytes()); t.put(p1); rs.getRegionServer().abort("I'm dead"); @@ -430,7 +430,7 @@ public class TestHCM { final Table table = connection.getTable(tableName); Put put = new Put(ROW); - put.add(FAM_NAM, ROW, ROW); + put.addColumn(FAM_NAM, ROW, ROW); table.put(put); ManualEnvironmentEdge mee = new ManualEnvironmentEdge(); @@ -583,7 +583,7 @@ public class TestHCM { TEST_UTIL.waitUntilAllRegionsAssigned(table.getName()); Put put = new Put(ROW); - put.add(FAM_NAM, ROW, ROW); + put.addColumn(FAM_NAM, ROW, ROW); table.put(put); ConnectionImplementation conn = (ConnectionImplementation) connection; @@ -606,7 +606,7 @@ public class TestHCM { conn.clearRegionCache(TABLE_NAME); Assert.assertEquals(0, conn.getNumberOfCachedRegionLocations(TABLE_NAME)); Put put2 = new Put(ROW); - put2.add(FAM_NAM, ROW, ROW); + put2.addColumn(FAM_NAM, ROW, ROW); table.put(put2); assertNotNull(conn.getCachedLocation(TABLE_NAME, ROW)); assertNotNull(conn.getCachedLocation(TableName.valueOf(TABLE_NAME.getName()), ROW.clone())); @@ -676,7 +676,7 @@ public class TestHCM { // We do a put and expect the cache to be updated, even if we don't retry LOG.info("Put starting"); Put put3 = new Put(ROW); - put3.add(FAM_NAM, ROW, ROW); + put3.addColumn(FAM_NAM, ROW, ROW); try { table.put(put3); Assert.fail("Unreachable point"); @@ -783,7 +783,7 @@ public class TestHCM { public void testCacheSeqNums() throws Exception{ Table table = TEST_UTIL.createMultiRegionTable(TABLE_NAME2, FAM_NAM); Put put = new Put(ROW); - put.add(FAM_NAM, ROW, ROW); + put.addColumn(FAM_NAM, ROW, ROW); table.put(put); ConnectionImplementation conn = (ConnectionImplementation) TEST_UTIL.getConnection(); @@ -913,7 +913,7 @@ public class TestHCM { } Put put = new Put(ROW_X); - put.add(FAM_NAM, ROW_X, ROW_X); + put.addColumn(FAM_NAM, ROW_X, ROW_X); table.put(put); // Now moving the region to the second server @@ -944,7 +944,7 @@ public class TestHCM { // If empty row, set it to first row.-f if (otherRow.length <= 0) otherRow = Bytes.toBytes("aaa"); Put put2 = new Put(otherRow); - put2.add(FAM_NAM, otherRow, otherRow); + put2.addColumn(FAM_NAM, otherRow, otherRow); table.put(put2); //cache put2's location // Check that we are in the expected state @@ -990,11 +990,11 @@ public class TestHCM { final int prevNumRetriesVal = setNumTries(conn, 2); Put put3 = new Put(ROW_X); - put3.add(FAM_NAM, ROW_X, ROW_X); + put3.addColumn(FAM_NAM, ROW_X, ROW_X); Put put4 = new Put(otherRow); - put4.add(FAM_NAM, otherRow, otherRow); + put4.addColumn(FAM_NAM, otherRow, otherRow); - // do multi + // do multi ArrayList actions = Lists.newArrayList(put4, put3); table.batch(actions, null); // first should be a valid row, // second we get RegionMovedException. diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java index 908fcdf..9aa493c 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java @@ -110,11 +110,11 @@ public class TestHTableMultiplexer { for (int i = 0; i < NUM_REGIONS; i++) { byte [] row = startRows[i]; if (row == null || row.length <= 0) continue; - Put put = new Put(row).add(FAMILY, QUALIFIER, VALUE1); + Put put = new Put(row).addColumn(FAMILY, QUALIFIER, VALUE1); success = multiplexer.put(TABLE_1, put); assertTrue("multiplexer.put returns", success); - put = new Put(row).add(FAMILY, QUALIFIER, VALUE1); + put = new Put(row).addColumn(FAMILY, QUALIFIER, VALUE1); success = multiplexer.put(TABLE_2, put); assertTrue("multiplexer.put failed", success); @@ -131,7 +131,7 @@ public class TestHTableMultiplexer { byte [] row = endRows[i]; if (row == null || row.length <= 0) continue; Put put = new Put(row); - put.add(FAMILY, QUALIFIER, VALUE2); + put.addColumn(FAMILY, QUALIFIER, VALUE2); multiput.add(put); } failedPuts = multiplexer.put(TABLE_1, multiput); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexerFlushCache.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexerFlushCache.java index b71e881..063e376 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexerFlushCache.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexerFlushCache.java @@ -97,7 +97,7 @@ public class TestHTableMultiplexerFlushCache { byte[] row = startRows[1]; assertTrue("2nd region should not start with empty row", row != null && row.length > 0); - Put put = new Put(row).add(FAMILY, QUALIFIER1, VALUE1); + Put put = new Put(row).addColumn(FAMILY, QUALIFIER1, VALUE1); assertTrue("multiplexer.put returns", multiplexer.put(TABLE, put)); checkExistence(htable, row, FAMILY, QUALIFIER1, VALUE1); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java index 0e5bd9c..6e088ab 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java @@ -201,7 +201,7 @@ public class TestMetaWithReplicas { htable = c.getTable(TABLE); byte[] row = "test".getBytes(); Put put = new Put(row); - put.add("foo".getBytes(), row, row); + put.addColumn("foo".getBytes(), row, row); BufferedMutator m = c.getBufferedMutator(TABLE); m.mutate(put); m.flush(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java index 21bdc5f..59c863e 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java @@ -213,10 +213,10 @@ public class TestMultiParallel { List actions = new ArrayList(); Put p = new Put(Bytes.toBytes("row1")); - p.add(Bytes.toBytes("bad_family"), Bytes.toBytes("qual"), Bytes.toBytes("value")); + p.addColumn(Bytes.toBytes("bad_family"), Bytes.toBytes("qual"), Bytes.toBytes("value")); actions.add(p); p = new Put(Bytes.toBytes("row2")); - p.add(BYTES_FAMILY, Bytes.toBytes("qual"), Bytes.toBytes("value")); + p.addColumn(BYTES_FAMILY, Bytes.toBytes("qual"), Bytes.toBytes("value")); actions.add(p); // row1 and row2 should be in the same region. @@ -423,7 +423,7 @@ public class TestMultiParallel { for (int i = 0; i < 100; i++) { Put put = new Put(ONE_ROW); byte[] qual = Bytes.toBytes("column" + i); - put.add(BYTES_FAMILY, qual, VALUE); + put.addColumn(BYTES_FAMILY, qual, VALUE); puts.add(put); } Object[] results = new Object[puts.size()]; @@ -464,8 +464,8 @@ public class TestMultiParallel { Delete d = new Delete(ONE_ROW); table.delete(d); Put put = new Put(ONE_ROW); - put.add(BYTES_FAMILY, QUAL1, Bytes.toBytes("abc")); - put.add(BYTES_FAMILY, QUAL2, Bytes.toBytes(1L)); + put.addColumn(BYTES_FAMILY, QUAL1, Bytes.toBytes("abc")); + put.addColumn(BYTES_FAMILY, QUAL2, Bytes.toBytes(1L)); table.put(put); Increment inc = new Increment(ONE_ROW); @@ -494,7 +494,7 @@ public class TestMultiParallel { final Connection connection = ConnectionFactory.createConnection(UTIL.getConfiguration()); Table table = connection.getTable(TEST_TABLE); Put put = new Put(ONE_ROW); - put.add(BYTES_FAMILY, QUALIFIER, Bytes.toBytes(0L)); + put.addColumn(BYTES_FAMILY, QUALIFIER, Bytes.toBytes(0L)); // Replace nonce manager with the one that returns each nonce twice. NonceGenerator cnm = new PerClientRandomNonceGenerator() { @@ -610,7 +610,7 @@ public class TestMultiParallel { // 2 put of new column Put put = new Put(KEYS[10]); - put.add(BYTES_FAMILY, qual2, val2); + put.addColumn(BYTES_FAMILY, qual2, val2); actions.add(put); // 3 delete @@ -629,7 +629,7 @@ public class TestMultiParallel { // 5 put of new column put = new Put(KEYS[40]); - put.add(BYTES_FAMILY, qual2, val2); + put.addColumn(BYTES_FAMILY, qual2, val2); actions.add(put); results = new Object[actions.size()]; @@ -673,7 +673,7 @@ public class TestMultiParallel { List puts = new ArrayList<>(); for (byte[] k : KEYS) { Put put = new Put(k); - put.add(BYTES_FAMILY, QUALIFIER, VALUE); + put.addColumn(BYTES_FAMILY, QUALIFIER, VALUE); puts.add(put); } return puts; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultipleTimestamps.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultipleTimestamps.java index bc62688..56f01c3 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultipleTimestamps.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultipleTimestamps.java @@ -475,8 +475,8 @@ public class TestMultipleTimestamps { for(int colIdx: columnIndexes) { byte column[] = Bytes.toBytes("column:" + colIdx); for (long version: versions) { - put.add(cf, column, version, Bytes.toBytes("value-version-" + - version)); + put.addColumn(cf, column, version, Bytes.toBytes("value-version-" + + version)); } } ht.put(put); @@ -496,7 +496,7 @@ public class TestMultipleTimestamps { put.setDurability(Durability.SKIP_WAL); for (long idx = versionStart; idx <= versionEnd; idx++) { - put.add(cf, column, idx, Bytes.toBytes("value-version-" + idx)); + put.addColumn(cf, column, idx, Bytes.toBytes("value-version-" + idx)); } ht.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutDeleteEtcCellIteration.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutDeleteEtcCellIteration.java index 9a8b166..60eb92f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutDeleteEtcCellIteration.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutDeleteEtcCellIteration.java @@ -49,7 +49,7 @@ public class TestPutDeleteEtcCellIteration { Put p = new Put(ROW); for (int i = 0; i < COUNT; i++) { byte [] bytes = Bytes.toBytes(i); - p.add(bytes, bytes, TIMESTAMP, bytes); + p.addColumn(bytes, bytes, TIMESTAMP, bytes); } int index = 0; for (CellScanner cellScanner = p.cellScanner(); cellScanner.advance();) { @@ -65,7 +65,7 @@ public class TestPutDeleteEtcCellIteration { Put p = new Put(ROW); for (int i = 0; i < COUNT; i++) { byte [] bytes = Bytes.toBytes(i); - p.add(bytes, bytes, TIMESTAMP, bytes); + p.addColumn(bytes, bytes, TIMESTAMP, bytes); } int index = 0; int trigger = 3; @@ -73,7 +73,7 @@ public class TestPutDeleteEtcCellIteration { Cell cell = cellScanner.current(); byte [] bytes = Bytes.toBytes(index++); // When we hit the trigger, try inserting a new KV; should trigger exception - if (trigger == 3) p.add(bytes, bytes, TIMESTAMP, bytes); + if (trigger == 3) p.addColumn(bytes, bytes, TIMESTAMP, bytes); cell.equals(new KeyValue(ROW, bytes, bytes, TIMESTAMP, bytes)); } assertEquals(COUNT, index); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutWithDelete.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutWithDelete.java index 0e819bb..0b5cd4e 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutWithDelete.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestPutWithDelete.java @@ -59,10 +59,10 @@ public class TestPutWithDelete { try { // put one row Put put = new Put(rowKey); - put.add(family, Bytes.toBytes("A"), Bytes.toBytes("a")); - put.add(family, Bytes.toBytes("B"), Bytes.toBytes("b")); - put.add(family, Bytes.toBytes("C"), Bytes.toBytes("c")); - put.add(family, Bytes.toBytes("D"), Bytes.toBytes("d")); + put.addColumn(family, Bytes.toBytes("A"), Bytes.toBytes("a")); + put.addColumn(family, Bytes.toBytes("B"), Bytes.toBytes("b")); + put.addColumn(family, Bytes.toBytes("C"), Bytes.toBytes("c")); + put.addColumn(family, Bytes.toBytes("D"), Bytes.toBytes("d")); table.put(put); // get row back and assert the values Get get = new Get(rowKey); @@ -77,11 +77,11 @@ public class TestPutWithDelete { Bytes.toString(result.getValue(family, Bytes.toBytes("D"))).equals("d")); // put the same row again with C column deleted put = new Put(rowKey); - put.add(family, Bytes.toBytes("A"), Bytes.toBytes("a1")); - put.add(family, Bytes.toBytes("B"), Bytes.toBytes("b1")); + put.addColumn(family, Bytes.toBytes("A"), Bytes.toBytes("a1")); + put.addColumn(family, Bytes.toBytes("B"), Bytes.toBytes("b1")); KeyValue marker = new KeyValue(rowKey, family, Bytes.toBytes("C"), HConstants.LATEST_TIMESTAMP, KeyValue.Type.DeleteColumn); - put.add(family, Bytes.toBytes("D"), Bytes.toBytes("d1")); + put.addColumn(family, Bytes.toBytes("D"), Bytes.toBytes("d1")); put.add(marker); table.put(put); // get row back and assert the values diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java index e6bde4e..b773b46 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java @@ -42,7 +42,6 @@ import org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.Bytes; -import org.apache.hadoop.hbase.util.FSTableDescriptors; import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster; import org.junit.AfterClass; @@ -197,7 +196,7 @@ public class TestReplicaWithCluster { bHdt.getColumnFamilies().length + 1, nHdt.getColumnFamilies().length); p = new Put(row); - p.add(row, row, row); + p.addColumn(row, row, row); table.put(p); g = new Get(row); @@ -254,7 +253,7 @@ public class TestReplicaWithCluster { admin.close(); Put p = new Put(row); - p.add(row, row, row); + p.addColumn(row, row, row); final Table table = HTU.getConnection().getTable(hdt.getTableName()); table.put(p); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java index a064bcc..d2e775d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java @@ -384,7 +384,7 @@ public class TestReplicasClient { flushRegion(hriSecondary); Put p = new Put(row); - p.add(f, row, row); + p.addColumn(f, row, row); table.put(p); flushRegion(hriPrimary); @@ -404,7 +404,7 @@ public class TestReplicasClient { flushRegion(hriPrimary); Put p = new Put(row); - p.add(f, row, row); + p.addColumn(f, row, row); table.put(p); flushRegion(hriPrimary); @@ -422,7 +422,7 @@ public class TestReplicasClient { flushRegion(hriSecondary); Put p = new Put(row); - p.add(f, row, row); + p.addColumn(f, row, row); table.put(p); flushRegion(hriSecondary); @@ -442,7 +442,7 @@ public class TestReplicasClient { try { // A simple put works, even if there here a second replica Put p = new Put(b1); - p.add(f, b1, b1); + p.addColumn(f, b1, b1); table.put(p); LOG.info("Put done"); @@ -534,12 +534,12 @@ public class TestReplicasClient { List puts = new ArrayList(2); byte[] b1 = Bytes.toBytes("testCancelOfMultiGet" + 0); Put p = new Put(b1); - p.add(f, b1, b1); + p.addColumn(f, b1, b1); puts.add(p); byte[] b2 = Bytes.toBytes("testCancelOfMultiGet" + 1); p = new Put(b2); - p.add(f, b2, b2); + p.addColumn(f, b2, b2); puts.add(p); table.put(puts); LOG.debug("PUT done"); @@ -619,7 +619,7 @@ public class TestReplicasClient { for (int i = 0; i < NUMROWS; i++) { byte[] b1 = Bytes.toBytes("testUseRegionWithReplica" + i); Put p = new Put(b1); - p.add(f, b1, b1); + p.addColumn(f, b1, b1); table.put(p); } LOG.debug("PUT done"); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRpcControllerFactory.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRpcControllerFactory.java index fca34c1..c087135 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRpcControllerFactory.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRpcControllerFactory.java @@ -135,7 +135,7 @@ public class TestRpcControllerFactory { Table table = connection.getTable(name); byte[] row = Bytes.toBytes("row"); Put p = new Put(row); - p.add(fam1, fam1, Bytes.toBytes("val0")); + p.addColumn(fam1, fam1, Bytes.toBytes("val0")); table.put(p); Integer counter = 1; @@ -147,7 +147,7 @@ public class TestRpcControllerFactory { counter = verifyCount(counter); Put p2 = new Put(row); - p2.add(fam1, Bytes.toBytes("qual"), Bytes.toBytes("val1")); + p2.addColumn(fam1, Bytes.toBytes("qual"), Bytes.toBytes("val1")); table.batch(Lists.newArrayList(p, p2), null); // this only goes to a single server, so we don't need to change the count here counter = verifyCount(counter); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannerTimeout.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannerTimeout.java index f86f248..6a9678e 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannerTimeout.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannerTimeout.java @@ -70,8 +70,8 @@ public class TestScannerTimeout { Table table = TEST_UTIL.createTable(TABLE_NAME, SOME_BYTES); for (int i = 0; i < NB_ROWS; i++) { Put put = new Put(Bytes.toBytes(i)); - put.add(SOME_BYTES, SOME_BYTES, SOME_BYTES); - table.put(put); + put.addColumn(SOME_BYTES, SOME_BYTES, SOME_BYTES); + table.put(put); } table.close(); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotCloneIndependence.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotCloneIndependence.java index 5f20790..18c17c7 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotCloneIndependence.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotCloneIndependence.java @@ -44,7 +44,6 @@ import org.junit.Before; import org.junit.BeforeClass; import org.junit.Ignore; import org.junit.Test; -import org.junit.Ignore; import org.junit.experimental.categories.Category; /** @@ -255,7 +254,7 @@ public class TestSnapshotCloneIndependence { final String rowKey = "new-row-" + System.currentTimeMillis(); Put p = new Put(Bytes.toBytes(rowKey)); - p.add(TEST_FAM, Bytes.toBytes("someQualifier"), Bytes.toBytes("someString")); + p.addColumn(TEST_FAM, Bytes.toBytes("someQualifier"), Bytes.toBytes("someString")); original.put(p); // Verify that it is not present in the original table @@ -266,7 +265,7 @@ public class TestSnapshotCloneIndependence { clonedTableRowCount, countRows(clonedTable)); p = new Put(Bytes.toBytes(rowKey)); - p.add(TEST_FAM, Bytes.toBytes("someQualifier"), Bytes.toBytes("someString")); + p.addColumn(TEST_FAM, Bytes.toBytes("someQualifier"), Bytes.toBytes("someString")); clonedTable.put(p); // Verify that the new family is not in the restored table's description diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java index 0692840..c803752 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java @@ -176,15 +176,15 @@ public class TestTimestampsFilter { Table ht = TEST_UTIL.createTable(TableName.valueOf(TABLE), FAMILIES, Integer.MAX_VALUE); Put p = new Put(Bytes.toBytes("row")); - p.add(FAMILY, Bytes.toBytes("column0"), 3, Bytes.toBytes("value0-3")); - p.add(FAMILY, Bytes.toBytes("column1"), 3, Bytes.toBytes("value1-3")); - p.add(FAMILY, Bytes.toBytes("column2"), 1, Bytes.toBytes("value2-1")); - p.add(FAMILY, Bytes.toBytes("column2"), 2, Bytes.toBytes("value2-2")); - p.add(FAMILY, Bytes.toBytes("column2"), 3, Bytes.toBytes("value2-3")); - p.add(FAMILY, Bytes.toBytes("column3"), 2, Bytes.toBytes("value3-2")); - p.add(FAMILY, Bytes.toBytes("column4"), 1, Bytes.toBytes("value4-1")); - p.add(FAMILY, Bytes.toBytes("column4"), 2, Bytes.toBytes("value4-2")); - p.add(FAMILY, Bytes.toBytes("column4"), 3, Bytes.toBytes("value4-3")); + p.addColumn(FAMILY, Bytes.toBytes("column0"), (long) 3, Bytes.toBytes("value0-3")); + p.addColumn(FAMILY, Bytes.toBytes("column1"), (long) 3, Bytes.toBytes("value1-3")); + p.addColumn(FAMILY, Bytes.toBytes("column2"), (long) 1, Bytes.toBytes("value2-1")); + p.addColumn(FAMILY, Bytes.toBytes("column2"), (long) 2, Bytes.toBytes("value2-2")); + p.addColumn(FAMILY, Bytes.toBytes("column2"), (long) 3, Bytes.toBytes("value2-3")); + p.addColumn(FAMILY, Bytes.toBytes("column3"), (long) 2, Bytes.toBytes("value3-2")); + p.addColumn(FAMILY, Bytes.toBytes("column4"), (long) 1, Bytes.toBytes("value4-1")); + p.addColumn(FAMILY, Bytes.toBytes("column4"), (long) 2, Bytes.toBytes("value4-2")); + p.addColumn(FAMILY, Bytes.toBytes("column4"), (long) 3, Bytes.toBytes("value4-3")); ht.put(p); ArrayList timestamps = new ArrayList(); @@ -360,7 +360,7 @@ public class TestTimestampsFilter { put.setDurability(Durability.SKIP_WAL); for (long idx = versionStart; idx <= versionEnd; idx++) { - put.add(cf, column, idx, Bytes.toBytes("value-version-" + idx)); + put.addColumn(cf, column, idx, Bytes.toBytes("value-version-" + idx)); } ht.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraint.java hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraint.java index 96da03a..0e47d39 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraint.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraint.java @@ -32,7 +32,6 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.MiscTests; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException; import org.apache.hadoop.hbase.client.Table; @@ -87,7 +86,8 @@ public class TestConstraint { // test that we don't fail on a valid put Put put = new Put(row1); byte[] value = Integer.toString(10).getBytes(); - put.add(dummy, new byte[0], value); + byte[] qualifier = new byte[0]; + put.addColumn(dummy, qualifier, value); table.put(put); } finally { table.close(); @@ -118,7 +118,8 @@ public class TestConstraint { // test that we do fail on violation Put put = new Put(row1); - put.add(dummy, new byte[0], "fail".getBytes()); + byte[] qualifier = new byte[0]; + put.addColumn(dummy, qualifier, "fail".getBytes()); LOG.warn("Doing put in table"); try { table.put(put); @@ -161,7 +162,8 @@ public class TestConstraint { try { // test that we don't fail because its disabled Put put = new Put(row1); - put.add(dummy, new byte[0], "pass".getBytes()); + byte[] qualifier = new byte[0]; + put.addColumn(dummy, qualifier, "pass".getBytes()); table.put(put); } finally { table.close(); @@ -193,7 +195,8 @@ public class TestConstraint { try { // test that we do fail on violation Put put = new Put(row1); - put.add(dummy, new byte[0], "pass".getBytes()); + byte[] qualifier = new byte[0]; + put.addColumn(dummy, qualifier, "pass".getBytes()); LOG.warn("Doing put in table"); table.put(put); } finally { @@ -225,8 +228,9 @@ public class TestConstraint { // test that we do fail on violation Put put = new Put(row1); - put.add(dummy, new byte[0], "pass".getBytes()); - + byte[] qualifier = new byte[0]; + put.addColumn(dummy, qualifier, "pass".getBytes()); + try{ table.put(put); fail("RuntimeFailConstraint wasn't triggered - this put shouldn't work!"); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestAggregateProtocol.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestAggregateProtocol.java index dfa0898..4fe0d23 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestAggregateProtocol.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestAggregateProtocol.java @@ -27,7 +27,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.Table; @@ -94,12 +93,12 @@ public class TestAggregateProtocol { Put put = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); Long l = new Long(i); - put.add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(l)); + put.addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(l)); table.put(put); Put p2 = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); - p2.add(TEST_FAMILY, Bytes.add(TEST_MULTI_CQ, Bytes.toBytes(l)), Bytes - .toBytes(l * 10)); + p2.addColumn(TEST_FAMILY, Bytes.add(TEST_MULTI_CQ, Bytes.toBytes(l)), Bytes + .toBytes(l * 10)); table.put(p2); } table.close(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java index a5ee4c1..d62e950 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestBatchCoprocessorEndpoint.java @@ -26,7 +26,6 @@ import java.util.Map; import java.util.TreeMap; import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.testclassification.CoprocessorTests; import org.apache.hadoop.hbase.testclassification.MediumTests; @@ -38,8 +37,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.coprocessor.Batch; import org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos; @@ -97,7 +94,7 @@ public class TestBatchCoprocessorEndpoint { Table table = util.getConnection().getTable(TEST_TABLE); for (int i = 0; i < ROWSIZE; i++) { Put put = new Put(ROWS[i]); - put.add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(i)); + put.addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(i)); table.put(put); } table.close(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestDoubleColumnInterpreter.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestDoubleColumnInterpreter.java index a1e67fe..ae3ed18 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestDoubleColumnInterpreter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestDoubleColumnInterpreter.java @@ -27,7 +27,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.Table; @@ -91,11 +90,11 @@ public class TestDoubleColumnInterpreter { Put put = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); Double d = new Double(i); - put.add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(d)); + put.addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(d)); table.put(put); Put p2 = new Put(ROWS[i]); put.setDurability(Durability.SKIP_WAL); - p2.add(TEST_FAMILY, Bytes.add(TEST_MULTI_CQ, Bytes.toBytes(d)), Bytes.toBytes(d * 0.10)); + p2.addColumn(TEST_FAMILY, Bytes.add(TEST_MULTI_CQ, Bytes.toBytes(d)), Bytes.toBytes(d * 0.10)); table.put(p2); } table.close(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java index c1d21fa..bd89744 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java @@ -108,15 +108,15 @@ public class TestHTableWrapper { table = util.createTable(TEST_TABLE, TEST_FAMILY); Put puta = new Put(ROW_A); - puta.add(TEST_FAMILY, qualifierCol1, bytes1); + puta.addColumn(TEST_FAMILY, qualifierCol1, bytes1); table.put(puta); Put putb = new Put(ROW_B); - putb.add(TEST_FAMILY, qualifierCol1, bytes2); + putb.addColumn(TEST_FAMILY, qualifierCol1, bytes2); table.put(putb); Put putc = new Put(ROW_C); - putc.add(TEST_FAMILY, qualifierCol1, bytes3); + putc.addColumn(TEST_FAMILY, qualifierCol1, bytes3); table.put(putc); } @@ -204,7 +204,7 @@ public class TestHTableWrapper { private void checkPutsAndDeletes() throws IOException { // put: - Put putD = new Put(ROW_D).add(TEST_FAMILY, qualifierCol1, bytes2); + Put putD = new Put(ROW_D).addColumn(TEST_FAMILY, qualifierCol1, bytes2); hTableInterface.put(putD); checkRowValue(ROW_D, bytes2); @@ -214,8 +214,8 @@ public class TestHTableWrapper { checkRowValue(ROW_D, null); // multiple puts: - Put[] puts = new Put[] { new Put(ROW_D).add(TEST_FAMILY, qualifierCol1, bytes2), - new Put(ROW_E).add(TEST_FAMILY, qualifierCol1, bytes3) }; + Put[] puts = new Put[] {new Put(ROW_D).addColumn(TEST_FAMILY, qualifierCol1, bytes2), + new Put(ROW_E).addColumn(TEST_FAMILY, qualifierCol1, bytes3)}; hTableInterface.put(Arrays.asList(puts)); checkRowsValues(new byte[][] { ROW_D, ROW_E }, new byte[][] { bytes2, bytes3 }); @@ -226,7 +226,7 @@ public class TestHTableWrapper { } private void checkCheckAndPut() throws IOException { - Put putC = new Put(ROW_C).add(TEST_FAMILY, qualifierCol1, bytes5); + Put putC = new Put(ROW_C).addColumn(TEST_FAMILY, qualifierCol1, bytes5); assertFalse(hTableInterface.checkAndPut(ROW_C, TEST_FAMILY, qualifierCol1, /* expect */bytes4, putC/* newValue */)); assertTrue(hTableInterface.checkAndPut(ROW_C, TEST_FAMILY, qualifierCol1, /* expect */bytes3, @@ -242,7 +242,7 @@ public class TestHTableWrapper { } private void checkIncrementColumnValue() throws IOException { - hTableInterface.put(new Put(ROW_A).add(TEST_FAMILY, qualifierCol1, Bytes.toBytes(1L))); + hTableInterface.put(new Put(ROW_A).addColumn(TEST_FAMILY, qualifierCol1, Bytes.toBytes(1L))); checkRowValue(ROW_A, Bytes.toBytes(1L)); final long newVal = hTableInterface @@ -319,7 +319,7 @@ public class TestHTableWrapper { } private void checkMutateRow() throws IOException { - Put put = new Put(ROW_A).add(TEST_FAMILY, qualifierCol1, bytes1); + Put put = new Put(ROW_A).addColumn(TEST_FAMILY, qualifierCol1, bytes1); RowMutations rowMutations = new RowMutations(ROW_A); rowMutations.add(put); hTableInterface.mutateRow(rowMutations); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java index 3122b4c..53b34b2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java @@ -102,7 +102,7 @@ public class TestOpenTableInCoprocessor { final WALEdit edit, final Durability durability) throws IOException { Table table = e.getEnvironment().getTable(otherTable, getPool()); Put p = new Put(new byte[] { 'a' }); - p.add(family, null, new byte[] { 'a' }); + p.addColumn(family, null, new byte[]{'a'}); try { table.batch(Collections.singletonList(put), null); } catch (InterruptedException e1) { @@ -162,7 +162,7 @@ public class TestOpenTableInCoprocessor { Table table = UTIL.getConnection().getTable(TableName.valueOf("primary")); Put p = new Put(new byte[] { 'a' }); - p.add(family, null, new byte[] { 'a' }); + p.addColumn(family, null, new byte[]{'a'}); table.put(p); table.close(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java index a02758dac..3cbbe9d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java @@ -33,7 +33,6 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Durability; @@ -95,7 +94,7 @@ public class TestRegionObserverBypass { public void testSimple() throws Exception { Table t = util.getConnection().getTable(tableName); Put p = new Put(row1); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); // before HBASE-4331, this would throw an exception t.put(p); checkRowAndDelete(t,row1,0); @@ -115,13 +114,13 @@ public class TestRegionObserverBypass { Table t = util.getConnection().getTable(tableName); List puts = new ArrayList(); Put p = new Put(row1); - p.add(dummy,dummy,dummy); + p.addColumn(dummy, dummy, dummy); puts.add(p); p = new Put(row2); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row3); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); // before HBASE-4331, this would throw an exception t.put(puts); @@ -131,13 +130,13 @@ public class TestRegionObserverBypass { puts.clear(); p = new Put(row1); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row2); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row3); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); // before HBASE-4331, this would throw an exception t.put(puts); @@ -147,13 +146,13 @@ public class TestRegionObserverBypass { puts.clear(); p = new Put(row1); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row2); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row3); - p.add(dummy,dummy,dummy); + p.addColumn(dummy, dummy, dummy); puts.add(p); // this worked fine even before HBASE-4331 t.put(puts); @@ -163,13 +162,13 @@ public class TestRegionObserverBypass { puts.clear(); p = new Put(row1); - p.add(dummy,dummy,dummy); + p.addColumn(dummy, dummy, dummy); puts.add(p); p = new Put(row2); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row3); - p.add(dummy,dummy,dummy); + p.addColumn(dummy, dummy, dummy); puts.add(p); // this worked fine even before HBASE-4331 t.put(puts); @@ -179,13 +178,13 @@ public class TestRegionObserverBypass { puts.clear(); p = new Put(row1); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); p = new Put(row2); - p.add(dummy,dummy,dummy); + p.addColumn(dummy, dummy, dummy); puts.add(p); p = new Put(row3); - p.add(test,dummy,dummy); + p.addColumn(test, dummy, dummy); puts.add(p); // before HBASE-4331, this would throw an exception t.put(puts); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java index c0f611a..abfadec 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java @@ -130,9 +130,9 @@ public class TestRegionObserverInterface { new Boolean[] { false, false, false, false, false, false, false, false }); Put put = new Put(ROW); - put.add(A, A, A); - put.add(B, B, B); - put.add(C, C, C); + put.addColumn(A, A, A); + put.addColumn(B, B, B); + put.addColumn(C, C, C); table.put(put); verifyMethodResult(SimpleRegionObserver.class, new String[] { "hadPreGet", "hadPostGet", @@ -189,9 +189,9 @@ public class TestRegionObserverInterface { tableName, new Boolean[] {false, false, false, false, false}); Put put = new Put(ROW); - put.add(A, A, A); - put.add(B, B, B); - put.add(C, C, C); + put.addColumn(A, A, A); + put.addColumn(B, B, B); + put.addColumn(C, C, C); Delete delete = new Delete(ROW); delete.addColumn(A, A); @@ -248,10 +248,10 @@ public class TestRegionObserverInterface { TableName.valueOf(TEST_TABLE.getNameAsString() + ".testCheckAndPutHooks"); try (Table table = util.createTable(tableName, new byte[][] {A, B, C})) { Put p = new Put(Bytes.toBytes(0)); - p.add(A, A, A); + p.addColumn(A, A, A); table.put(p); p = new Put(Bytes.toBytes(0)); - p.add(A, A, A); + p.addColumn(A, A, A); verifyMethodResult(SimpleRegionObserver.class, new String[] { "hadPreCheckAndPut", "hadPreCheckAndPutAfterRowLock", "hadPostCheckAndPut" }, tableName, new Boolean[] { false, false, false }); @@ -274,7 +274,7 @@ public class TestRegionObserverInterface { Table table = util.createTable(tableName, new byte[][] {A, B, C}); try { Put p = new Put(Bytes.toBytes(0)); - p.add(A, A, A); + p.addColumn(A, A, A); table.put(p); Delete d = new Delete(Bytes.toBytes(0)); table.delete(d); @@ -338,7 +338,7 @@ public class TestRegionObserverInterface { Table table = util.getConnection().getTable(tableName); Put put = new Put(ROW); - put.add(A, A, A); + put.addColumn(A, A, A); table.put(put); Get get = new Get(ROW); @@ -412,7 +412,7 @@ public class TestRegionObserverInterface { Table table = util.getConnection().getTable(tableName); Put put = new Put(ROW); - put.add(A, A, A); + put.addColumn(A, A, A); table.put(put); Delete delete = new Delete(ROW); @@ -522,7 +522,7 @@ public class TestRegionObserverInterface { byte[] iBytes = Bytes.toBytes(i); Put put = new Put(iBytes); put.setDurability(Durability.SKIP_WAL); - put.add(A, A, iBytes); + put.addColumn(A, A, iBytes); table.put(put); } @@ -677,9 +677,9 @@ public class TestRegionObserverInterface { } Put put = new Put(ROW); - put.add(A, A, A); - put.add(B, B, B); - put.add(C, C, C); + put.addColumn(A, A, A); + put.addColumn(B, B, B); + put.addColumn(C, C, C); table.put(put); verifyMethodResult(SimpleRegionObserver.Legacy.class, @@ -729,9 +729,9 @@ public class TestRegionObserverInterface { } Put put = new Put(ROW); - put.add(A, A, A); - put.add(B, B, B); - put.add(C, C, C); + put.addColumn(A, A, A); + put.addColumn(B, B, B); + put.addColumn(C, C, C); table.put(put); cluster.killRegionServer(rs1.getRegionServer().getServerName()); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java index 00808bd..e87eaec 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java @@ -177,7 +177,7 @@ public class TestRegionObserverScannerOpenHook { h.load(EmptyRegionObsever.class, Coprocessor.PRIORITY_USER, conf); Put put = new Put(ROW); - put.add(A, A, A); + put.addColumn(A, A, A); region.put(put); Get get = new Get(ROW); @@ -203,7 +203,7 @@ public class TestRegionObserverScannerOpenHook { // put a row and flush it to disk Put put = new Put(ROW); - put.add(A, A, A); + put.addColumn(A, A, A); region.put(put); region.flush(true); Get get = new Get(ROW); @@ -269,7 +269,7 @@ public class TestRegionObserverScannerOpenHook { // put a row and flush it to disk Put put = new Put(ROW); - put.add(A, A, A); + put.addColumn(A, A, A); table.put(put); HRegionServer rs = UTIL.getRSForFirstRegionInTable(desc.getTableName()); @@ -282,7 +282,7 @@ public class TestRegionObserverScannerOpenHook { // put another row and flush that too put = new Put(Bytes.toBytes("anotherrow")); - put.add(A, A, A); + put.addColumn(A, A, A); table.put(put); admin.flushRegion(region.getRegionInfo().getRegionName()); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverStacking.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverStacking.java index 0a4ca16..723edcb 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverStacking.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverStacking.java @@ -127,7 +127,7 @@ public class TestRegionObserverStacking extends TestCase { h.load(ObserverC.class, Coprocessor.PRIORITY_LOWEST, conf); Put put = new Put(ROW); - put.add(A, A, A); + put.addColumn(A, A, A); region.put(put); Coprocessor c = h.findCoprocessor(ObserverA.class.getName()); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorExceptionWithAbort.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorExceptionWithAbort.java index ef62564..d25948b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorExceptionWithAbort.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorExceptionWithAbort.java @@ -33,7 +33,6 @@ import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.Waiter.Predicate; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.regionserver.HRegionServer; @@ -108,7 +107,7 @@ public class TestRegionServerCoprocessorExceptionWithAbort { try { final byte[] ROW = Bytes.toBytes("aaa"); Put put = new Put(ROW); - put.add(TEST_FAMILY, ROW, ROW); + put.addColumn(TEST_FAMILY, ROW, ROW); table.put(put); } catch (IOException e) { // The region server is going to be aborted. diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java index 5f2b7bd..5d4721a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java @@ -138,15 +138,15 @@ public class TestRowProcessorEndpoint { table = util.createTable(TABLE, FAM); { Put put = new Put(ROW); - put.add(FAM, A, Bytes.add(B, C)); // B, C are friends of A - put.add(FAM, B, Bytes.add(D, E, F)); // D, E, F are friends of B - put.add(FAM, C, G); // G is a friend of C + put.addColumn(FAM, A, Bytes.add(B, C)); + put.addColumn(FAM, B, Bytes.add(D, E, F)); + put.addColumn(FAM, C, G); table.put(put); rowSize = put.size(); } Put put = new Put(ROW2); - put.add(FAM, D, E); - put.add(FAM, F, G); + put.addColumn(FAM, D, E); + put.addColumn(FAM, F, G); table.put(put); row2Size = put.size(); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java index 7772664..75fe7a2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java @@ -489,7 +489,7 @@ public class TestWALObserver { private Put creatPutWith2Families(byte[] row) throws IOException { Put p = new Put(row); for (int i = 0; i < TEST_FAMILY.length - 1; i++) { - p.add(TEST_FAMILY[i], TEST_QUALIFIER[i], TEST_VALUE[i]); + p.addColumn(TEST_FAMILY[i], TEST_QUALIFIER[i], TEST_VALUE[i]); } return p; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestDependentColumnFilter.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestDependentColumnFilter.java index 6ccfc62..3a635b8 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestDependentColumnFilter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestDependentColumnFilter.java @@ -100,26 +100,26 @@ public class TestDependentColumnFilter { private void addData() throws IOException { Put put = new Put(ROWS[0]); // add in an entry for each stamp, with 2 as a "good" value - put.add(FAMILIES[0], QUALIFIER, STAMPS[0], BAD_VALS[0]); - put.add(FAMILIES[0], QUALIFIER, STAMPS[1], BAD_VALS[1]); - put.add(FAMILIES[0], QUALIFIER, STAMPS[2], MATCH_VAL); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[0], BAD_VALS[0]); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[1], BAD_VALS[1]); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[2], MATCH_VAL); // add in entries for stamps 0 and 2. // without a value check both will be "accepted" // with one 2 will be accepted(since the corresponding ts entry // has a matching value - put.add(FAMILIES[1], QUALIFIER, STAMPS[0], BAD_VALS[0]); - put.add(FAMILIES[1], QUALIFIER, STAMPS[2], BAD_VALS[2]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[0], BAD_VALS[0]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[2], BAD_VALS[2]); this.region.put(put); put = new Put(ROWS[1]); - put.add(FAMILIES[0], QUALIFIER, STAMPS[0], BAD_VALS[0]); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[0], BAD_VALS[0]); // there is no corresponding timestamp for this so it should never pass - put.add(FAMILIES[0], QUALIFIER, STAMPS[2], MATCH_VAL); + put.addColumn(FAMILIES[0], QUALIFIER, STAMPS[2], MATCH_VAL); // if we reverse the qualifiers this one should pass - put.add(FAMILIES[1], QUALIFIER, STAMPS[0], MATCH_VAL); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[0], MATCH_VAL); // should pass - put.add(FAMILIES[1], QUALIFIER, STAMPS[1], BAD_VALS[2]); + put.addColumn(FAMILIES[1], QUALIFIER, STAMPS[1], BAD_VALS[2]); this.region.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilter.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilter.java index 5825a4f..8fcec9f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilter.java @@ -152,7 +152,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_ONE) { - p.add(FAMILIES[0], QUALIFIER, VALUES[0]); + p.addColumn(FAMILIES[0], QUALIFIER, VALUES[0]); } this.region.put(p); } @@ -160,7 +160,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_TWO) { - p.add(FAMILIES[1], QUALIFIER, VALUES[1]); + p.addColumn(FAMILIES[1], QUALIFIER, VALUES[1]); } this.region.put(p); } @@ -173,7 +173,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_ONE) { - p.add(FAMILIES[1], QUALIFIER, VALUES[0]); + p.addColumn(FAMILIES[1], QUALIFIER, VALUES[0]); } this.region.put(p); } @@ -181,7 +181,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for(byte [] QUALIFIER : QUALIFIERS_TWO) { - p.add(FAMILIES[0], QUALIFIER, VALUES[1]); + p.addColumn(FAMILIES[0], QUALIFIER, VALUES[1]); } this.region.put(p); } @@ -229,7 +229,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for (byte[] QUALIFIER : QUALIFIERS_THREE) { - p.add(FAMILIES[0], QUALIFIER, VALUES[0]); + p.addColumn(FAMILIES[0], QUALIFIER, VALUES[0]); } this.region.put(p); @@ -238,7 +238,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for (byte[] QUALIFIER : QUALIFIERS_FOUR) { - p.add(FAMILIES[1], QUALIFIER, VALUES[1]); + p.addColumn(FAMILIES[1], QUALIFIER, VALUES[1]); } this.region.put(p); } @@ -250,7 +250,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for (byte[] QUALIFIER : QUALIFIERS_THREE) { - p.add(FAMILIES[1], QUALIFIER, VALUES[0]); + p.addColumn(FAMILIES[1], QUALIFIER, VALUES[0]); } this.region.put(p); } @@ -258,7 +258,7 @@ public class TestFilter { Put p = new Put(ROW); p.setDurability(Durability.SKIP_WAL); for (byte[] QUALIFIER : QUALIFIERS_FOUR) { - p.add(FAMILIES[0], QUALIFIER, VALUES[1]); + p.addColumn(FAMILIES[0], QUALIFIER, VALUES[1]); } this.region.put(p); } @@ -1458,7 +1458,7 @@ public class TestFilter { for(int i=0; i<5; i++) { Put p = new Put(Bytes.toBytes((char)('a'+i) + "row")); p.setDurability(Durability.SKIP_WAL); - p.add(family, qualifier, Bytes.toBytes(String.valueOf(111+i))); + p.addColumn(family, qualifier, Bytes.toBytes(String.valueOf(111 + i))); testRegion.put(p); } testRegion.flush(true); @@ -1501,7 +1501,7 @@ public class TestFilter { // Need to change one of the group one columns to use group two value Put p = new Put(ROWS_ONE[2]); - p.add(FAMILIES[0], QUALIFIERS_ONE[2], VALUES[1]); + p.addColumn(FAMILIES[0], QUALIFIERS_ONE[2], VALUES[1]); this.region.put(p); // Now let's grab rows that have Q_ONE[0](VALUES[0]) and Q_ONE[2](VALUES[1]) @@ -1819,8 +1819,8 @@ public class TestFilter { // Test that the filter skips multiple column versions. Put p = new Put(ROWS_ONE[0]); p.setDurability(Durability.SKIP_WAL); - p.add(FAMILIES[0], QUALIFIERS_ONE[0], VALUES[0]); - this.region.put(p); + p.addColumn(FAMILIES[0], QUALIFIERS_ONE[0], VALUES[0]); + this.region.put(p); this.region.flush(true); // Set of KVs (page: 1; pageSize: 1) - the first set of 1 column per row @@ -2017,7 +2017,7 @@ public class TestFilter { for(int i=0; i<10; i++) { Put p = new Put(Bytes.toBytes("row" + i)); p.setDurability(Durability.SKIP_WAL); - p.add(FAMILIES[0], columnStatus, Bytes.toBytes(i%2)); + p.addColumn(FAMILIES[0], columnStatus, Bytes.toBytes(i % 2)); testRegion.put(p); } testRegion.flush(true); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java index 78a4d1f..0d2940c 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java @@ -105,8 +105,7 @@ public class TestFilterWithScanLimits extends FilterTestingCluster { for (int i = 1; i < 4; i++) { Put put = new Put(Bytes.toBytes("row" + i)); for (int j = 1; j < 6; j++) { - put.add(Bytes.toBytes("f1"), Bytes.toBytes("c" + j), - Bytes.toBytes(i + "_c" + j)); + put.addColumn(Bytes.toBytes("f1"), Bytes.toBytes("c" + j), Bytes.toBytes(i + "_c" + j)); } puts.add(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java index 8ce0b76..d30790a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java @@ -39,7 +39,6 @@ import org.apache.hadoop.hbase.ZooKeeperConnectionException; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -130,8 +129,7 @@ public class TestFilterWrapper { long timestamp = j; if (i != 1) timestamp = i; - put.add(Bytes.toBytes("f1"), Bytes.toBytes("c" + j), timestamp, - Bytes.toBytes(i + "_c" + j)); + put.addColumn(Bytes.toBytes("f1"), Bytes.toBytes("c" + j), timestamp, Bytes.toBytes(i + "_c" + j)); } puts.add(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java index 680ce5c..145e42f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java @@ -116,7 +116,7 @@ public class TestFuzzyRowAndColumnRangeFilter { Put p = new Put(rk); p.setDurability(Durability.SKIP_WAL); - p.add(cf.getBytes(), cq, Bytes.toBytes(c)); + p.addColumn(cf.getBytes(), cq, Bytes.toBytes(c)); ht.put(p); LOG.info("Inserting: rk: " + Bytes.toStringBinary(rk) + " cq: " + Bytes.toStringBinary(cq)); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java index 5c78dfe..ba1d2a1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java @@ -139,7 +139,7 @@ public class TestFuzzyRowFilterEndToEnd { Put p = new Put(rk); p.setDurability(Durability.SKIP_WAL); - p.add(cf.getBytes(), cq, Bytes.toBytes(c)); + p.addColumn(cf.getBytes(), cq, Bytes.toBytes(c)); ht.put(p); } } @@ -277,7 +277,7 @@ public class TestFuzzyRowFilterEndToEnd { Put p = new Put(rk); p.setDurability(Durability.SKIP_WAL); - p.add(cf.getBytes(), cq, Bytes.toBytes(c)); + p.addColumn(cf.getBytes(), cq, Bytes.toBytes(c)); ht.put(p); LOG.info("Inserting: rk: " + Bytes.toStringBinary(rk) + " cq: " + Bytes.toStringBinary(cq)); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestInvocationRecordFilter.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestInvocationRecordFilter.java index a8651d8..dd763b7 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestInvocationRecordFilter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestInvocationRecordFilter.java @@ -75,8 +75,7 @@ public class TestInvocationRecordFilter { Put put = new Put(ROW_BYTES); for (int i = 0; i < 10; i += 2) { // puts 0, 2, 4, 6 and 8 - put.add(FAMILY_NAME_BYTES, Bytes.toBytes(QUALIFIER_PREFIX + i), i, - Bytes.toBytes(VALUE_PREFIX + i)); + put.addColumn(FAMILY_NAME_BYTES, Bytes.toBytes(QUALIFIER_PREFIX + i), (long) i, Bytes.toBytes(VALUE_PREFIX + i)); } this.region.put(put); this.region.flush(true); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java index 3be10ec..21a0df1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java @@ -75,7 +75,7 @@ public class TestScanRowPrefix extends FilterTestingCluster { for (byte[] rowId: rowIds) { Put p = new Put(rowId); // Use the rowId as the column qualifier - p.add("F".getBytes(), rowId, "Dummy value".getBytes()); + p.addColumn("F".getBytes(), rowId, "Dummy value".getBytes()); table.put(p); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/fs/TestBlockReorder.java hbase-server/src/test/java/org/apache/hadoop/hbase/fs/TestBlockReorder.java index 504350c..c529107 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/fs/TestBlockReorder.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/fs/TestBlockReorder.java @@ -318,7 +318,7 @@ public class TestBlockReorder { // insert one put to ensure a minimal size Put p = new Put(sb); - p.add(sb, sb, sb); + p.addColumn(sb, sb, sb); h.put(p); DirectoryListing dl = dfs.getClient().listPaths(rootDir, HdfsFileStatus.EMPTY_NAME); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java index 53deeb2..d7e555d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java @@ -133,8 +133,7 @@ public class TestChangingEncoding { for (int i = 0; i < NUM_ROWS_PER_BATCH; ++i) { Put put = new Put(getRowKey(batchId, i)); for (int j = 0; j < NUM_COLS_PER_ROW; ++j) { - put.add(CF_BYTES, getQualifier(j), - getValue(batchId, i, j)); + put.addColumn(CF_BYTES, getQualifier(j), getValue(batchId, i, j)); } put.setDurability(Durability.SKIP_WAL); puts.add(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java index 5ccb206..ce66e82 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java @@ -152,7 +152,7 @@ public class TestEncodedSeekers { KeyValue kv = new KeyValue(key, CF_BYTES, col, HConstants.LATEST_TIMESTAMP, value, tag); put.add(kv); } else { - put.add(CF_BYTES, col, value); + put.addColumn(CF_BYTES, col, value); } if(VERBOSE){ KeyValue kvPut = new KeyValue(key, CF_BYTES, col, value); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java index cf2aca5..9eb4734 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java @@ -131,8 +131,7 @@ public class TestForceCacheImportantBlocks { Put put = new Put(Bytes.toBytes("row" + i)); for (int j = 0; j < NUM_COLS_PER_ROW; ++j) { for (long ts = 1; ts < NUM_TIMESTAMPS_PER_COL; ++ts) { - put.add(CF_BYTES, Bytes.toBytes("col" + j), ts, - Bytes.toBytes("value" + i + "_" + j + "_" + ts)); + put.addColumn(CF_BYTES, Bytes.toBytes("col" + j), ts, Bytes.toBytes("value" + i + "_" + j + "_" + ts)); } } region.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingKeyRange.java hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingKeyRange.java index 7584cf2..de5e918 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingKeyRange.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingKeyRange.java @@ -106,8 +106,7 @@ public class TestScannerSelectionUsingKeyRange { for (int iRow = 0; iRow < NUM_ROWS; ++iRow) { Put put = new Put(Bytes.toBytes("row" + iRow)); for (int iCol = 0; iCol < NUM_COLS_PER_ROW; ++iCol) { - put.add(FAMILY_BYTES, Bytes.toBytes("col" + iCol), - Bytes.toBytes("value" + iFile + "_" + iRow + "_" + iCol)); + put.addColumn(FAMILY_BYTES, Bytes.toBytes("col" + iCol), Bytes.toBytes("value" + iFile + "_" + iRow + "_" + iCol)); } region.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingTTL.java hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingTTL.java index d5f4bcd..0234dd8 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingTTL.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerSelectionUsingTTL.java @@ -120,8 +120,7 @@ public class TestScannerSelectionUsingTTL { for (int iRow = 0; iRow < NUM_ROWS; ++iRow) { Put put = new Put(Bytes.toBytes("row" + iRow)); for (int iCol = 0; iCol < NUM_COLS_PER_ROW; ++iCol) { - put.add(FAMILY_BYTES, Bytes.toBytes("col" + iCol), - ts + version, Bytes.toBytes("value" + iFile + "_" + iRow + "_" + iCol)); + put.addColumn(FAMILY_BYTES, Bytes.toBytes("col" + iCol), ts + version, Bytes.toBytes("value" + iFile + "_" + iRow + "_" + iCol)); } region.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java index e0b7a0b..5be0e72 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java @@ -38,7 +38,6 @@ import org.apache.hadoop.hbase.*; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -127,12 +126,12 @@ public class TestTableInputFormat { Table table = UTIL.createTable(TableName.valueOf(tableName), families); Put p = new Put("aaa".getBytes()); for (byte[] family : families) { - p.add(family, null, "value aaa".getBytes()); + p.addColumn(family, null, "value aaa".getBytes()); } table.put(p); p = new Put("bbb".getBytes()); for (byte[] family : families) { - p.add(family, null, "value bbb".getBytes()); + p.addColumn(family, null, "value bbb".getBytes()); } table.put(p); return table; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduceUtil.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduceUtil.java index 628bb96..3080bc3 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduceUtil.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduceUtil.java @@ -116,8 +116,7 @@ public class TestTableMapReduceUtil { for (String president : presidentsRowKeys) { if (presidentNames.hasNext()) { Put p = new Put(Bytes.toBytes(president)); - p.add(COLUMN_FAMILY, COLUMN_QUALIFIER, - Bytes.toBytes(presidentNames.next())); + p.addColumn(COLUMN_FAMILY, COLUMN_QUALIFIER, Bytes.toBytes(presidentNames.next())); table.put(p); } } @@ -125,7 +124,7 @@ public class TestTableMapReduceUtil { for (String actor : actorsRowKeys) { if (actorNames.hasNext()) { Put p = new Put(Bytes.toBytes(actor)); - p.add(COLUMN_FAMILY, COLUMN_QUALIFIER, Bytes.toBytes(actorNames.next())); + p.addColumn(COLUMN_FAMILY, COLUMN_QUALIFIER, Bytes.toBytes(actorNames.next())); table.put(p); } } @@ -267,8 +266,7 @@ public class TestTableMapReduceUtil { String name = Bytes.toString(result.getValue(COLUMN_FAMILY, COLUMN_QUALIFIER)); - outCollector.collect(outKey, new Put(Bytes.toBytes("rowKey2")).add( - COLUMN_FAMILY, COLUMN_QUALIFIER, Bytes.toBytes(name))); + outCollector.collect(outKey, new Put(Bytes.toBytes("rowKey2")).addColumn(COLUMN_FAMILY, COLUMN_QUALIFIER, Bytes.toBytes(name))); } } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCellCounter.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCellCounter.java index 2d3d6ef..507f2ab 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCellCounter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCellCounter.java @@ -32,7 +32,6 @@ import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.MapReduceTests; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.LauncherSecurityManager; -import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.util.ToolRunner; import org.junit.AfterClass; import org.junit.BeforeClass; @@ -86,14 +85,14 @@ public class TestCellCounter { Table t = UTIL.createTable(sourceTable, families); try{ Put p = new Put(ROW1); - p.add(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); - p.add(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); - p.add(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); + p.addColumn(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); + p.addColumn(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); + p.addColumn(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); t.put(p); p = new Put(ROW2); - p.add(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); - p.add(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); - p.add(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); + p.addColumn(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); + p.addColumn(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); + p.addColumn(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); t.put(p); String[] args = { sourceTable.getNameAsString(), FQ_OUTPUT_DIR.toString(), ";", "^row1" }; runCount(args); @@ -124,15 +123,15 @@ public class TestCellCounter { Table t = UTIL.createTable(sourceTable, families); try{ Put p = new Put(ROW1); - p.add(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); - p.add(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); - p.add(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); - t.put(p); + p.addColumn(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); + p.addColumn(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); + p.addColumn(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); + t.put(p); p = new Put(ROW2); - p.add(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); - p.add(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); - p.add(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); - t.put(p); + p.addColumn(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); + p.addColumn(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); + p.addColumn(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); + t.put(p); String[] args = { sourceTable.getNameAsString(), FQ_OUTPUT_DIR.toString(), ";", "^row1", "--starttime=" + now, "--endtime=" + now + 2 }; @@ -164,15 +163,15 @@ public class TestCellCounter { Table t = UTIL.createTable(sourceTable, families); try{ Put p = new Put(ROW1); - p.add(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); - p.add(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); - p.add(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); - t.put(p); + p.addColumn(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); + p.addColumn(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); + p.addColumn(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); + t.put(p); p = new Put(ROW2); - p.add(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); - p.add(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); - p.add(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); - t.put(p); + p.addColumn(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); + p.addColumn(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); + p.addColumn(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); + t.put(p); String[] args = { sourceTable.getNameAsString(), FQ_OUTPUT_DIR.toString(), ";", "^row1", "--endtime=" + now + 1 }; @@ -204,15 +203,15 @@ public class TestCellCounter { Table t = UTIL.createTable(sourceTable, families); try{ Put p = new Put(ROW1); - p.add(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); - p.add(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); - p.add(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); - t.put(p); + p.addColumn(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); + p.addColumn(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); + p.addColumn(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); + t.put(p); p = new Put(ROW2); - p.add(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); - p.add(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); - p.add(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); - t.put(p); + p.addColumn(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); + p.addColumn(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); + p.addColumn(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); + t.put(p); String[] args = { sourceTable.getNameAsString(), FQ_OUTPUT_DIR.toString(), ";", "--starttime=" + now + 1, "--endtime=" + now + 2 }; @@ -286,14 +285,14 @@ public class TestCellCounter { Table t = UTIL.createTable(sourceTable, families); try { Put p = new Put(ROW1); - p.add(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); - p.add(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); - p.add(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); + p.addColumn(FAMILY_A, QUALIFIER, now, Bytes.toBytes("Data11")); + p.addColumn(FAMILY_B, QUALIFIER, now + 1, Bytes.toBytes("Data12")); + p.addColumn(FAMILY_A, QUALIFIER, now + 2, Bytes.toBytes("Data13")); t.put(p); p = new Put(ROW2); - p.add(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); - p.add(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); - p.add(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); + p.addColumn(FAMILY_B, QUALIFIER, now, Bytes.toBytes("Dat21")); + p.addColumn(FAMILY_A, QUALIFIER, now + 1, Bytes.toBytes("Data22")); + p.addColumn(FAMILY_B, QUALIFIER, now + 2, Bytes.toBytes("Data23")); t.put(p); String[] args = { sourceTable.getNameAsString(), outputDir.toString(), ";" }; runCount(args); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java index 118395b..dc27177 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java @@ -24,7 +24,6 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; import java.io.ByteArrayOutputStream; -import java.io.IOException; import java.io.PrintStream; import org.apache.hadoop.conf.Configuration; @@ -39,7 +38,6 @@ import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.MapReduceTests; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.LauncherSecurityManager; -import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.util.ToolRunner; import org.junit.AfterClass; import org.junit.BeforeClass; @@ -84,7 +82,7 @@ public class TestCopyTable { // put rows into the first table for (int i = 0; i < 10; i++) { Put p = new Put(Bytes.toBytes("row" + i)); - p.add(FAMILY, COLUMN1, COLUMN1); + p.addColumn(FAMILY, COLUMN1, COLUMN1); t1.put(p); } @@ -147,13 +145,13 @@ public class TestCopyTable { // put rows into the first table Put p = new Put(ROW0); - p.add(FAMILY, COLUMN1, COLUMN1); + p.addColumn(FAMILY, COLUMN1, COLUMN1); t1.put(p); p = new Put(ROW1); - p.add(FAMILY, COLUMN1, COLUMN1); + p.addColumn(FAMILY, COLUMN1, COLUMN1); t1.put(p); p = new Put(ROW2); - p.add(FAMILY, COLUMN1, COLUMN1); + p.addColumn(FAMILY, COLUMN1, COLUMN1); t1.put(p); CopyTable copy = new CopyTable(); @@ -197,14 +195,14 @@ public class TestCopyTable { Table t = TEST_UTIL.createTable(sourceTable, families); Table t2 = TEST_UTIL.createTable(targetTable, families); Put p = new Put(ROW1); - p.add(FAMILY_A, QUALIFIER, Bytes.toBytes("Data11")); - p.add(FAMILY_B, QUALIFIER, Bytes.toBytes("Data12")); - p.add(FAMILY_A, QUALIFIER, Bytes.toBytes("Data13")); + p.addColumn(FAMILY_A, QUALIFIER, Bytes.toBytes("Data11")); + p.addColumn(FAMILY_B, QUALIFIER, Bytes.toBytes("Data12")); + p.addColumn(FAMILY_A, QUALIFIER, Bytes.toBytes("Data13")); t.put(p); p = new Put(ROW2); - p.add(FAMILY_B, QUALIFIER, Bytes.toBytes("Dat21")); - p.add(FAMILY_A, QUALIFIER, Bytes.toBytes("Data22")); - p.add(FAMILY_B, QUALIFIER, Bytes.toBytes("Data23")); + p.addColumn(FAMILY_B, QUALIFIER, Bytes.toBytes("Dat21")); + p.addColumn(FAMILY_A, QUALIFIER, Bytes.toBytes("Data22")); + p.addColumn(FAMILY_B, QUALIFIER, Bytes.toBytes("Data23")); t.put(p); long currentTime = System.currentTimeMillis(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java index 90b92bb..04e3266 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java @@ -1006,7 +1006,7 @@ public class TestHFileOutputFormat { // put some data in it and flush to create a storefile Put p = new Put(Bytes.toBytes("test")); - p.add(FAMILIES[0], Bytes.toBytes("1"), Bytes.toBytes("1")); + p.addColumn(FAMILIES[0], Bytes.toBytes("1"), Bytes.toBytes("1")); table.put(p); admin.flush(TABLE_NAME); assertEquals(1, util.countRows(table)); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java index 7a13e5d..11b1f51 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java @@ -1046,7 +1046,7 @@ public class TestHFileOutputFormat2 { // put some data in it and flush to create a storefile Put p = new Put(Bytes.toBytes("test")); - p.add(FAMILIES[0], Bytes.toBytes("1"), Bytes.toBytes("1")); + p.addColumn(FAMILIES[0], Bytes.toBytes("1"), Bytes.toBytes("1")); table.put(p); admin.flush(TABLE_NAME); assertEquals(1, util.countRows(table)); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java index 29a4538..cf7a092 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java @@ -71,7 +71,6 @@ import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.VerySlowMapReduceTests; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.LauncherSecurityManager; -import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.Mapper.Context; import org.apache.hadoop.util.ToolRunner; import org.junit.After; @@ -174,14 +173,14 @@ public class TestImportExport { String EXPORT_TABLE = "exportSimpleCase"; Table t = UTIL.createTable(TableName.valueOf(EXPORT_TABLE), FAMILYA, 3); Put p = new Put(ROW1); - p.add(FAMILYA, QUAL, now, QUAL); - p.add(FAMILYA, QUAL, now+1, QUAL); - p.add(FAMILYA, QUAL, now+2, QUAL); + p.addColumn(FAMILYA, QUAL, now, QUAL); + p.addColumn(FAMILYA, QUAL, now + 1, QUAL); + p.addColumn(FAMILYA, QUAL, now + 2, QUAL); t.put(p); p = new Put(ROW2); - p.add(FAMILYA, QUAL, now, QUAL); - p.add(FAMILYA, QUAL, now+1, QUAL); - p.add(FAMILYA, QUAL, now+2, QUAL); + p.addColumn(FAMILYA, QUAL, now, QUAL); + p.addColumn(FAMILYA, QUAL, now + 1, QUAL); + p.addColumn(FAMILYA, QUAL, now + 2, QUAL); t.put(p); String[] args = new String[] { @@ -274,12 +273,12 @@ public class TestImportExport { Table t = UTIL.getConnection().getTable(desc.getTableName()); Put p = new Put(ROW1); - p.add(FAMILYA, QUAL, now, QUAL); - p.add(FAMILYA, QUAL, now+1, QUAL); - p.add(FAMILYA, QUAL, now+2, QUAL); - p.add(FAMILYA, QUAL, now+3, QUAL); - p.add(FAMILYA, QUAL, now+4, QUAL); - t.put(p); + p.addColumn(FAMILYA, QUAL, now, QUAL); + p.addColumn(FAMILYA, QUAL, now + 1, QUAL); + p.addColumn(FAMILYA, QUAL, now + 2, QUAL); + p.addColumn(FAMILYA, QUAL, now + 3, QUAL); + p.addColumn(FAMILYA, QUAL, now + 4, QUAL); + t.put(p); String[] args = new String[] { "-D" + Export.EXPORT_BATCHING + "=" + EXPORT_BATCH_SIZE, // added scanner batching arg. @@ -305,11 +304,11 @@ public class TestImportExport { Table t = UTIL.getConnection().getTable(desc.getTableName()); Put p = new Put(ROW1); - p.add(FAMILYA, QUAL, now, QUAL); - p.add(FAMILYA, QUAL, now+1, QUAL); - p.add(FAMILYA, QUAL, now+2, QUAL); - p.add(FAMILYA, QUAL, now+3, QUAL); - p.add(FAMILYA, QUAL, now+4, QUAL); + p.addColumn(FAMILYA, QUAL, now, QUAL); + p.addColumn(FAMILYA, QUAL, now + 1, QUAL); + p.addColumn(FAMILYA, QUAL, now + 2, QUAL); + p.addColumn(FAMILYA, QUAL, now + 3, QUAL); + p.addColumn(FAMILYA, QUAL, now + 4, QUAL); t.put(p); Delete d = new Delete(ROW1, now+3); @@ -373,7 +372,7 @@ public class TestImportExport { //Add first version of QUAL Put p = new Put(ROW1); - p.add(FAMILYA, QUAL, now, QUAL); + p.addColumn(FAMILYA, QUAL, now, QUAL); exportT.put(p); //Add Delete family marker @@ -382,7 +381,7 @@ public class TestImportExport { //Add second version of QUAL p = new Put(ROW1); - p.add(FAMILYA, QUAL, now+5, "s".getBytes()); + p.addColumn(FAMILYA, QUAL, now + 5, "s".getBytes()); exportT.put(p); //Add second Delete family marker @@ -449,15 +448,15 @@ public class TestImportExport { Table exportTable = UTIL.getConnection().getTable(desc.getTableName()); Put p1 = new Put(ROW1); - p1.add(FAMILYA, QUAL, now, QUAL); - p1.add(FAMILYA, QUAL, now + 1, QUAL); - p1.add(FAMILYA, QUAL, now + 2, QUAL); - p1.add(FAMILYA, QUAL, now + 3, QUAL); - p1.add(FAMILYA, QUAL, now + 4, QUAL); + p1.addColumn(FAMILYA, QUAL, now, QUAL); + p1.addColumn(FAMILYA, QUAL, now + 1, QUAL); + p1.addColumn(FAMILYA, QUAL, now + 2, QUAL); + p1.addColumn(FAMILYA, QUAL, now + 3, QUAL); + p1.addColumn(FAMILYA, QUAL, now + 4, QUAL); // Having another row would actually test the filter. Put p2 = new Put(ROW2); - p2.add(FAMILYA, QUAL, now, QUAL); + p2.addColumn(FAMILYA, QUAL, now, QUAL); exportTable.put(Arrays.asList(p1, p2)); @@ -641,15 +640,15 @@ public class TestImportExport { // Insert some data Put put = new Put(ROW1); - put.add(FAMILYA, QUAL, now, QUAL); - put.add(FAMILYA, QUAL, now + 1, QUAL); - put.add(FAMILYA, QUAL, now + 2, QUAL); + put.addColumn(FAMILYA, QUAL, now, QUAL); + put.addColumn(FAMILYA, QUAL, now + 1, QUAL); + put.addColumn(FAMILYA, QUAL, now + 2, QUAL); exportTable.put(put); put = new Put(ROW2); - put.add(FAMILYA, QUAL, now, QUAL); - put.add(FAMILYA, QUAL, now + 1, QUAL); - put.add(FAMILYA, QUAL, now + 2, QUAL); + put.addColumn(FAMILYA, QUAL, now, QUAL); + put.addColumn(FAMILYA, QUAL, now + 1, QUAL); + put.addColumn(FAMILYA, QUAL, now + 2, QUAL); exportTable.put(put); // Run the export diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java index 70420e5..60d1dfa 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java @@ -34,9 +34,7 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.HBaseTestingUtility; -import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -119,7 +117,7 @@ public class TestMultithreadedTableMapper { newValue.reverse(); // Now set the value to be collected Put outval = new Put(key.get()); - outval.add(OUTPUT_FAMILY, null, Bytes.toBytes(newValue.toString())); + outval.addColumn(OUTPUT_FAMILY, null, Bytes.toBytes(newValue.toString())); context.write(key, outval); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRowCounter.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRowCounter.java index 2024429..4a9f2c4 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRowCounter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRowCounter.java @@ -25,21 +25,16 @@ import static org.junit.Assert.fail; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.PrintStream; -import java.sql.Time; import java.util.ArrayList; -import java.util.Arrays; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.testclassification.MapReduceTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; -import org.apache.hadoop.hbase.mapreduce.RowCounter.RowCounterMapper; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.LauncherSecurityManager; import org.apache.hadoop.util.ToolRunner; @@ -157,13 +152,13 @@ public class TestRowCounter { // clean up content of TABLE_NAME Table table = TEST_UTIL.deleteTableData(TableName.valueOf(TABLE_NAME)); ts = System.currentTimeMillis(); - put1.add(family, col1, ts, Bytes.toBytes("val1")); + put1.addColumn(family, col1, ts, Bytes.toBytes("val1")); table.put(put1); Thread.sleep(100); ts = System.currentTimeMillis(); - put2.add(family, col1, ts, Bytes.toBytes("val2")); - put3.add(family, col1, ts, Bytes.toBytes("val3")); + put2.addColumn(family, col1, ts, Bytes.toBytes("val2")); + put3.addColumn(family, col1, ts, Bytes.toBytes("val3")); table.put(put2); table.put(put3); table.close(); @@ -229,9 +224,9 @@ public class TestRowCounter { for (; i < TOTAL_ROWS - ROWS_WITH_ONE_COL; i++) { byte[] row = Bytes.toBytes("row" + i); Put put = new Put(row); - put.add(family, col1, value); - put.add(family, col2, value); - put.add(family, col3, value); + put.addColumn(family, col1, value); + put.addColumn(family, col2, value); + put.addColumn(family, col3, value); rowsUpdate.add(put); } @@ -239,7 +234,7 @@ public class TestRowCounter { for (; i < TOTAL_ROWS; i++) { byte[] row = Bytes.toBytes("row" + i); Put put = new Put(row); - put.add(family, col2, value); + put.addColumn(family, col2, value); rowsUpdate.add(put); } table.put(rowsUpdate); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java index eb25972..bbc446b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java @@ -37,7 +37,6 @@ import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.*; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -57,7 +56,6 @@ import org.apache.hadoop.mapred.MiniMRCluster; import org.apache.hadoop.mapreduce.InputFormat; import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.JobContext; -import org.apache.hadoop.mapreduce.Mapper.Context; import org.apache.hadoop.mapreduce.lib.output.NullOutputFormat; import org.junit.AfterClass; import org.junit.Before; @@ -123,12 +121,12 @@ public class TestTableInputFormat { Table table = UTIL.createTable(TableName.valueOf(tableName), families); Put p = new Put("aaa".getBytes()); for (byte[] family : families) { - p.add(family, null, "value aaa".getBytes()); + p.addColumn(family, null, "value aaa".getBytes()); } table.put(p); p = new Put("bbb".getBytes()); for (byte[] family : families) { - p.add(family, null, "value bbb".getBytes()); + p.addColumn(family, null, "value bbb".getBytes()); } table.put(p); return table; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java index 1dcbd2a..a52eea6 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java @@ -29,7 +29,6 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; @@ -86,7 +85,7 @@ public class TestTableMapReduce extends TestTableMapReduceBase { newValue.reverse(); // Now set the value to be collected Put outval = new Put(key.get()); - outval.add(OUTPUT_FAMILY, null, Bytes.toBytes(newValue.toString())); + outval.addColumn(OUTPUT_FAMILY, null, Bytes.toBytes(newValue.toString())); context.write(key, outval); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java index 572d9ae..b662928 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java @@ -33,7 +33,6 @@ import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -129,7 +128,7 @@ public abstract class TestTableMapReduceBase { // Now set the value to be collected Put outval = new Put(key.get()); - outval.add(OUTPUT_FAMILY, null, Bytes.toBytes(newValue.toString())); + outval.addColumn(OUTPUT_FAMILY, null, Bytes.toBytes(newValue.toString())); return outval; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java index 03da1ed..1a3f566 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java @@ -121,7 +121,7 @@ public class TestTimeRangeMapRed { for (Long ts : tsList) { Put put = new Put(key.get()); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_NAME, COLUMN_NAME, ts, Bytes.toBytes(true)); + put.addColumn(FAMILY_NAME, COLUMN_NAME, ts, Bytes.toBytes(true)); puts.add(put); } table.put(puts); @@ -156,7 +156,7 @@ public class TestTimeRangeMapRed { for (Map.Entry entry : TIMESTAMP.entrySet()) { Put put = new Put(KEY); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_NAME, COLUMN_NAME, entry.getKey(), Bytes.toBytes(false)); + put.addColumn(FAMILY_NAME, COLUMN_NAME, entry.getKey(), Bytes.toBytes(false)); puts.add(put); } Table table = UTIL.getConnection().getTable(desc.getTableName()); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestWALPlayer.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestWALPlayer.java index 1b99818..112e263 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestWALPlayer.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestWALPlayer.java @@ -99,8 +99,8 @@ public class TestWALPlayer { // put a row into the first table Put p = new Put(ROW); - p.add(FAMILY, COLUMN1, COLUMN1); - p.add(FAMILY, COLUMN2, COLUMN2); + p.addColumn(FAMILY, COLUMN1, COLUMN1); + p.addColumn(FAMILY, COLUMN2, COLUMN2); t1.put(p); // delete one column Delete d = new Delete(ROW); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java index abeab3f..5b3abea 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java @@ -208,7 +208,7 @@ public class TestAssignmentListener { for (int i = 0; i < 10; ++i) { byte[] key = Bytes.toBytes("row-" + i); Put put = new Put(key); - put.add(FAMILY, null, key); + put.addColumn(FAMILY, null, key); table.put(put); } } finally { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java index 4a43bbd..21af809 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java @@ -968,7 +968,7 @@ public class TestDistributedLogSplitting { key = new byte[] { 0, 0, 0, 0, 1 }; } Put put = new Put(key); - put.add(Bytes.toBytes("family"), Bytes.toBytes("c1"), new byte[]{'b'}); + put.addColumn(Bytes.toBytes("family"), Bytes.toBytes("c1"), new byte[]{'b'}); ht.put(put); } catch (IOException ioe) { Assert.assertTrue(ioe instanceof RetriesExhaustedWithDetailsException); @@ -1668,7 +1668,7 @@ public class TestDistributedLogSplitting { for(int i = 0; i < numRows; i++) { Put put = new Put(Bytes.add(startRow, Bytes.toBytes(i))); for(byte [] family : families) { - put.add(family, qf, null); + put.addColumn(family, qf, null); } region.put(put); } @@ -1689,7 +1689,7 @@ public class TestDistributedLogSplitting { k[1] = b2; k[2] = b3; Put put = new Put(k); - put.add(f, column, k); + put.addColumn(f, column, k); puts.add(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestGetLastFlushedSequenceId.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestGetLastFlushedSequenceId.java index d2bb11a..a7e161b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestGetLastFlushedSequenceId.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestGetLastFlushedSequenceId.java @@ -29,7 +29,6 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.NamespaceDescriptor; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds; @@ -74,7 +73,7 @@ public class TestGetLastFlushedSequenceId { testUtil.getHBaseAdmin().createNamespace( NamespaceDescriptor.create(tableName.getNamespaceAsString()).build()); Table table = testUtil.createTable(tableName, families); - table.put(new Put(Bytes.toBytes("k")).add(family, Bytes.toBytes("q"), Bytes.toBytes("v"))); + table.put(new Put(Bytes.toBytes("k")).addColumn(family, Bytes.toBytes("q"), Bytes.toBytes("v"))); MiniHBaseCluster cluster = testUtil.getMiniHBaseCluster(); List rsts = cluster.getRegionServerThreads(); Region region = null; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java index 7cea0df..4e6460b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java @@ -27,7 +27,6 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Result; @@ -504,7 +503,7 @@ public class TestMasterTransitions { byte [] row = getStartKey(hri); Put p = new Put(row); p.setDurability(Durability.SKIP_WAL); - p.add(getTestFamily(), getTestQualifier(), row); + p.addColumn(getTestFamily(), getTestQualifier(), row); t.put(p); rows++; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java index ed5d919..53ee92b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java @@ -93,7 +93,7 @@ public class TestWarmupRegion { for (int i = 0; i < numRows; i++) { long ts = System.currentTimeMillis() * 2; Put put = new Put(ROW, ts); - put.add(FAMILY, COLUMN, VALUE); + put.addColumn(FAMILY, COLUMN, VALUE); table.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizerOnCluster.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizerOnCluster.java index 2cf26c0..d24023d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizerOnCluster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizerOnCluster.java @@ -210,7 +210,7 @@ public class TestSimpleRegionNormalizerOnCluster { Put put = new Put(key); byte[] col = Bytes.toBytes(String.valueOf(j)); byte[] value = dataGenerator.generateRandomSizeValue(key, col); - put.add(FAMILYNAME, col, value); + put.addColumn(FAMILYNAME, col, value); region.put(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java index b3f29db..45b62cd 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java @@ -420,7 +420,7 @@ public class MasterProcedureTestingUtility { Put put = new Put(key); put.setDurability(Durability.SKIP_WAL); for (byte[] family: families) { - put.add(family, q, value); + put.addColumn(family, q, value); } return put; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java index 6d2dd3e..0c06588 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java @@ -26,10 +26,8 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.Put; -import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.security.User; @@ -40,7 +38,6 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper; import org.apache.hadoop.hbase.util.ManualEnvironmentEdge; import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread; -import org.apache.hadoop.security.UserGroupInformation; import org.junit.After; import org.junit.AfterClass; @@ -511,7 +508,7 @@ public class TestQuotaThrottle { try { while (count < maxOps) { Put put = new Put(Bytes.toBytes("row-" + count)); - put.add(FAMILY, QUALIFIER, Bytes.toBytes("data-" + count)); + put.addColumn(FAMILY, QUALIFIER, Bytes.toBytes("data-" + count)); for (final Table table: tables) { table.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestAtomicOperation.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestAtomicOperation.java index 3edbc08..5ea219a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestAtomicOperation.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestAtomicOperation.java @@ -386,7 +386,7 @@ public class TestAtomicOperation { RowMutations rm = new RowMutations(row); if (op) { Put p = new Put(row, ts); - p.add(fam1, qual1, value1); + p.addColumn(fam1, qual1, value1); p.setDurability(Durability.ASYNC_WAL); rm.add(p); Delete d = new Delete(row); @@ -399,7 +399,7 @@ public class TestAtomicOperation { d.setDurability(Durability.ASYNC_WAL); rm.add(d); Put p = new Put(row, ts); - p.add(fam1, qual2, value2); + p.addColumn(fam1, qual2, value2); p.setDurability(Durability.ASYNC_WAL); rm.add(p); } @@ -479,7 +479,7 @@ public class TestAtomicOperation { List mrm = new ArrayList(); if (op) { Put p = new Put(row2, ts); - p.add(fam1, qual1, value1); + p.addColumn(fam1, qual1, value1); p.setDurability(Durability.ASYNC_WAL); mrm.add(p); Delete d = new Delete(row); @@ -493,7 +493,7 @@ public class TestAtomicOperation { mrm.add(d); Put p = new Put(row, ts); p.setDurability(Durability.ASYNC_WAL); - p.add(fam1, qual1, value2); + p.addColumn(fam1, qual1, value2); mrm.add(p); } region.mutateRowsWithLocks(mrm, rowsToLock, HConstants.NO_NONCE, HConstants.NO_NONCE); @@ -581,7 +581,7 @@ public class TestAtomicOperation { Put[] puts = new Put[1]; Put put = new Put(Bytes.toBytes("r1")); - put.add(Bytes.toBytes(family), Bytes.toBytes("q1"), Bytes.toBytes("10")); + put.addColumn(Bytes.toBytes(family), Bytes.toBytes("q1"), Bytes.toBytes("10")); puts[0] = put; region.batchMutate(puts, HConstants.NO_NONCE, HConstants.NO_NONCE); @@ -615,7 +615,7 @@ public class TestAtomicOperation { public void doWork() throws Exception { Put[] puts = new Put[1]; Put put = new Put(Bytes.toBytes("r1")); - put.add(Bytes.toBytes(family), Bytes.toBytes("q1"), Bytes.toBytes("50")); + put.addColumn(Bytes.toBytes(family), Bytes.toBytes("q1"), Bytes.toBytes("50")); puts[0] = put; testStep = TestStep.PUT_STARTED; region.batchMutate(puts, HConstants.NO_NONCE, HConstants.NO_NONCE); @@ -632,7 +632,7 @@ public class TestAtomicOperation { public void doWork() throws Exception { Put[] puts = new Put[1]; Put put = new Put(Bytes.toBytes("r1")); - put.add(Bytes.toBytes(family), Bytes.toBytes("q1"), Bytes.toBytes("11")); + put.addColumn(Bytes.toBytes(family), Bytes.toBytes("q1"), Bytes.toBytes("11")); puts[0] = put; while (testStep != TestStep.PUT_COMPLETED) { Thread.sleep(100); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBlocksRead.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBlocksRead.java index 2d1507b..2dfbee6 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBlocksRead.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBlocksRead.java @@ -127,7 +127,7 @@ public class TestBlocksRead { put.setDurability(Durability.SKIP_WAL); for (long version = versionStart; version <= versionEnd; version++) { - put.add(cf, columnBytes, version, genValue(row, col, version)); + put.addColumn(cf, columnBytes, version, genValue(row, col, version)); } region.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java index 49d32d2..3e5b25c 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java @@ -151,7 +151,7 @@ public class TestCompaction { Put p = new Put(Bytes.add(STARTROW, Bytes.toBytes(i))); p.setDurability(Durability.SKIP_WAL); for (int j = 0; j < jmax; j++) { - p.add(COLUMN_FAMILY, Bytes.toBytes(j), pad); + p.addColumn(COLUMN_FAMILY, Bytes.toBytes(j), pad); } HBaseTestCase.addContent(loader, Bytes.toString(COLUMN_FAMILY)); loader.put(p); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java index f809a97..22b2163 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java @@ -222,7 +222,7 @@ public class TestCompactionState { byte[] row = Bytes.toBytes(random.nextLong()); Put p = new Put(row); for (int j = 0; j < families.length; ++j) { - p.add(families[ j ], qualifier, row); + p.addColumn(families[j], qualifier, row); } puts.add(p); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCorruptedRegionStoreFile.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCorruptedRegionStoreFile.java index 7d1a39c..54dbe9b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCorruptedRegionStoreFile.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCorruptedRegionStoreFile.java @@ -88,7 +88,7 @@ public class TestCorruptedRegionStoreFile { while (rowCount < NUM_ROWS) { Put put = new Put(Bytes.toBytes(String.format("%010d", rowCount))); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_NAME, q, value); + put.addColumn(FAMILY_NAME, q, value); table.put(put); if ((rowCount++ % ROW_PER_FILE) == 0) { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java index 45a95c4..f195663 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java @@ -215,8 +215,7 @@ public class TestEncryptionKeyRotation { // Create a store file Table table = TEST_UTIL.getConnection().getTable(htd.getTableName()); try { - table.put(new Put(Bytes.toBytes("testrow")) - .add(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value"))); + table.put(new Put(Bytes.toBytes("testrow")).addColumn(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value"))); } finally { table.close(); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java index 0a6b2b5..ba2c3a2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java @@ -106,8 +106,7 @@ public class TestEncryptionRandomKeying { // Create a store file Table table = TEST_UTIL.getConnection().getTable(htd.getTableName()); try { - table.put(new Put(Bytes.toBytes("testrow")) - .add(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value"))); + table.put(new Put(Bytes.toBytes("testrow")).addColumn(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value"))); } finally { table.close(); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java index dbc8d15..b5ea863 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java @@ -282,13 +282,13 @@ public class TestHRegion { byte [] value = Bytes.toBytes(name.getMethodName()); // Make a random put against our cf. Put put = new Put(value); - put.add(COLUMN_FAMILY_BYTES, null, value); + put.addColumn(COLUMN_FAMILY_BYTES, null, value); // First put something in current memstore, which will be in snapshot after flusher.prepare() region.put(put); StoreFlushContext storeFlushCtx = store.createFlushContext(12345); storeFlushCtx.prepare(); // Second put something in current memstore - put.add(COLUMN_FAMILY_BYTES, Bytes.toBytes("abc"), value); + put.addColumn(COLUMN_FAMILY_BYTES, Bytes.toBytes("abc"), value); region.put(put); // Close with something in memstore and something in the snapshot. Make sure all is cleared. region.close(); @@ -334,7 +334,7 @@ public class TestHRegion { faultyLog.setStoreFlushCtx(store.createFlushContext(12345)); Put put = new Put(value); - put.add(COLUMN_FAMILY_BYTES, Bytes.toBytes("abc"), value); + put.addColumn(COLUMN_FAMILY_BYTES, Bytes.toBytes("abc"), value); faultyLog.setFailureType(FaultyFSLog.FailureType.SYNC); boolean threwIOE = false; @@ -383,7 +383,7 @@ public class TestHRegion { // Put some value and make sure flush could be completed normally byte [] value = Bytes.toBytes(name.getMethodName()); Put put = new Put(value); - put.add(COLUMN_FAMILY_BYTES, Bytes.toBytes("abc"), value); + put.addColumn(COLUMN_FAMILY_BYTES, Bytes.toBytes("abc"), value); region.put(put); long onePutSize = region.getMemstoreSize(); assertTrue(onePutSize > 0); @@ -452,7 +452,7 @@ public class TestHRegion { Assert.assertEquals(0, size); // Put one item into memstore. Measure the size of one item in memstore. Put p1 = new Put(row); - p1.add(new KeyValue(row, COLUMN_FAMILY_BYTES, qual1, 1, (byte[])null)); + p1.add(new KeyValue(row, COLUMN_FAMILY_BYTES, qual1, 1, (byte[]) null)); region.put(p1); final long sizeOfOnePut = region.getMemstoreSize(); // Fail a flush which means the current memstore will hang out as memstore 'snapshot'. @@ -551,7 +551,7 @@ public class TestHRegion { this.region = initHRegion(tableName, method, CONF, family); Put put = new Put(Bytes.toBytes("r1")); - put.add(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); + put.addColumn(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); region.put(put); region.flush(true); @@ -599,10 +599,10 @@ public class TestHRegion { this.region = initHRegion(tableName, method, CONF, family); Put put = new Put(Bytes.toBytes("r1")); - put.add(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); + put.addColumn(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); region.put(put); put = new Put(Bytes.toBytes("r2")); - put.add(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); + put.addColumn(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); region.put(put); region.flush(true); @@ -848,7 +848,7 @@ public class TestHRegion { for (long i = minSeqId; i < maxSeqId; i++) { Put put = new Put(Bytes.toBytes(i)); - put.add(family, Bytes.toBytes(i), Bytes.toBytes(i)); + put.addColumn(family, Bytes.toBytes(i), Bytes.toBytes(i)); region.put(put); region.flush(true); } @@ -950,7 +950,7 @@ public class TestHRegion { for (long i = minSeqId; i < maxSeqId; i++) { Put put = new Put(Bytes.toBytes(i)); - put.add(family, Bytes.toBytes(i), Bytes.toBytes(i)); + put.addColumn(family, Bytes.toBytes(i), Bytes.toBytes(i)); region.put(put); region.flush(true); } @@ -1147,7 +1147,7 @@ public class TestHRegion { int i = 0; Put put = new Put(Bytes.toBytes(i)); put.setDurability(Durability.SKIP_WAL); // have to skip mocked wal - put.add(family, Bytes.toBytes(i), Bytes.toBytes(i)); + put.addColumn(family, Bytes.toBytes(i), Bytes.toBytes(i)); region.put(put); // 1. Test case where START_FLUSH throws exception @@ -1435,11 +1435,11 @@ public class TestHRegion { System.out.println(String.format("Saving row: %s, with value %s", row, value)); Put put = new Put(Bytes.toBytes(row)); put.setDurability(Durability.SKIP_WAL); - put.add(Bytes.toBytes("trans-blob"), null, Bytes.toBytes("value for blob")); - put.add(Bytes.toBytes("trans-type"), null, Bytes.toBytes("statement")); - put.add(Bytes.toBytes("trans-date"), null, Bytes.toBytes("20090921010101999")); - put.add(Bytes.toBytes("trans-tags"), Bytes.toBytes("qual2"), Bytes.toBytes(value)); - put.add(Bytes.toBytes("trans-group"), null, Bytes.toBytes("adhocTransactionGroupId")); + put.addColumn(Bytes.toBytes("trans-blob"), null, Bytes.toBytes("value for blob")); + put.addColumn(Bytes.toBytes("trans-type"), null, Bytes.toBytes("statement")); + put.addColumn(Bytes.toBytes("trans-date"), null, Bytes.toBytes("20090921010101999")); + put.addColumn(Bytes.toBytes("trans-tags"), Bytes.toBytes("qual2"), Bytes.toBytes(value)); + put.addColumn(Bytes.toBytes("trans-group"), null, Bytes.toBytes("adhocTransactionGroupId")); r.put(put); } } @@ -1452,7 +1452,7 @@ public class TestHRegion { try { Put p = new Put(b.toBytes()); byte[] cfwithcolon = Bytes.toBytes(COLUMN_FAMILY + ":"); - p.add(cfwithcolon, cfwithcolon, cfwithcolon); + p.addColumn(cfwithcolon, cfwithcolon, cfwithcolon); boolean exception = false; try { this.region.put(p); @@ -1481,7 +1481,7 @@ public class TestHRegion { final Put[] puts = new Put[10]; for (int i = 0; i < 10; i++) { puts[i] = new Put(Bytes.toBytes("row_" + i)); - puts[i].add(cf, qual, val); + puts[i].addColumn(cf, qual, val); } OperationStatus[] codes = this.region.batchMutate(puts); @@ -1492,7 +1492,7 @@ public class TestHRegion { metricsAssertHelper.assertCounter("syncTimeNumOps", syncs + 1, source); LOG.info("Next a batch put with one invalid family"); - puts[5].add(Bytes.toBytes("BAD_CF"), qual, val); + puts[5].addColumn(Bytes.toBytes("BAD_CF"), qual, val); codes = this.region.batchMutate(puts); assertEquals(10, codes.length); for (int i = 0; i < 10; i++) { @@ -1521,9 +1521,9 @@ public class TestHRegion { final Put[] puts = new Put[10]; for (int i = 0; i < 10; i++) { puts[i] = new Put(Bytes.toBytes("row_" + i)); - puts[i].add(cf, qual, val); + puts[i].addColumn(cf, qual, val); } - puts[5].add(Bytes.toBytes("BAD_CF"), qual, val); + puts[5].addColumn(Bytes.toBytes("BAD_CF"), qual, val); LOG.info("batchPut will have to break into four batches to avoid row locks"); RowLock rowLock1 = region.getRowLock(Bytes.toBytes("row_2")); @@ -1628,7 +1628,7 @@ public class TestHRegion { final Put[] puts = new Put[10]; for (int i = 0; i < 10; i++) { puts[i] = new Put(Bytes.toBytes("row_" + i), Long.MAX_VALUE - 100); - puts[i].add(cf, qual, val); + puts[i].addColumn(cf, qual, val); } OperationStatus[] codes = this.region.batchMutate(puts); @@ -1663,7 +1663,7 @@ public class TestHRegion { try { // Putting empty data in key Put put = new Put(row1); - put.add(fam1, qf1, emptyVal); + put.addColumn(fam1, qf1, emptyVal); // checkAndPut with empty value boolean res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL, new BinaryComparator( @@ -1672,7 +1672,7 @@ public class TestHRegion { // Putting data in key put = new Put(row1); - put.add(fam1, qf1, val1); + put.addColumn(fam1, qf1, val1); // checkAndPut with correct value res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL, new BinaryComparator(emptyVal), @@ -1691,7 +1691,7 @@ public class TestHRegion { assertFalse(res); put = new Put(row1); - put.add(fam1, qf1, val2); + put.addColumn(fam1, qf1, val2); // checkAndPut with correct value res = region.checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL, new BinaryComparator(val1), put, true); @@ -1712,7 +1712,7 @@ public class TestHRegion { // checkAndPut looking for a null value put = new Put(row1); - put.add(fam1, qf1, val1); + put.addColumn(fam1, qf1, val1); res = region .checkAndMutate(row1, fam1, qf1, CompareOp.EQUAL, new NullComparator(), put, true); @@ -1737,7 +1737,7 @@ public class TestHRegion { try { // Putting data in key Put put = new Put(row1); - put.add(fam1, qf1, val1); + put.addColumn(fam1, qf1, val1); region.put(put); // checkAndPut with wrong value @@ -1770,7 +1770,7 @@ public class TestHRegion { try { // Putting data in key Put put = new Put(row1); - put.add(fam1, qf1, val1); + put.addColumn(fam1, qf1, val1); region.put(put); // checkAndPut with correct value @@ -1806,7 +1806,7 @@ public class TestHRegion { try { // Putting val3 in key Put put = new Put(row1); - put.add(fam1, qf1, val3); + put.addColumn(fam1, qf1, val3); region.put(put); // Test CompareOp.LESS: original = val3, compare with val3, fail @@ -1822,7 +1822,7 @@ public class TestHRegion { // Test CompareOp.LESS: original = val3, compare with val2, // succeed (now value = val2) put = new Put(row1); - put.add(fam1, qf1, val2); + put.addColumn(fam1, qf1, val2); res = region.checkAndMutate(row1, fam1, qf1, CompareOp.LESS, new BinaryComparator(val2), put, true); assertEquals(true, res); @@ -1841,7 +1841,7 @@ public class TestHRegion { // Test CompareOp.LESS_OR_EQUAL: original = val2, compare with val1, // succeed (now value = val3) put = new Put(row1); - put.add(fam1, qf1, val3); + put.addColumn(fam1, qf1, val3); res = region.checkAndMutate(row1, fam1, qf1, CompareOp.LESS_OR_EQUAL, new BinaryComparator(val1), put, true); assertEquals(true, res); @@ -1859,7 +1859,7 @@ public class TestHRegion { // Test CompareOp.GREATER: original = val3, compare with val4, // succeed (now value = val2) put = new Put(row1); - put.add(fam1, qf1, val2); + put.addColumn(fam1, qf1, val2); res = region.checkAndMutate(row1, fam1, qf1, CompareOp.GREATER, new BinaryComparator(val4), put, true); assertEquals(true, res); @@ -1902,7 +1902,7 @@ public class TestHRegion { try { // Putting data in the key to check Put put = new Put(row1); - put.add(fam1, qf1, val1); + put.addColumn(fam1, qf1, val1); region.put(put); // Creating put to add @@ -1938,7 +1938,7 @@ public class TestHRegion { this.region = initHRegion(tableName, this.getName(), CONF, COLUMNS); try { Put put = new Put(row2); - put.add(fam1, qual1, value1); + put.addColumn(fam1, qual1, value1); try { region.checkAndMutate(row, fam1, qual1, CompareOp.EQUAL, new BinaryComparator(value2), put, false); @@ -1973,16 +1973,16 @@ public class TestHRegion { try { // Put content Put put = new Put(row1); - put.add(fam1, qf1, val1); + put.addColumn(fam1, qf1, val1); region.put(put); Threads.sleep(2); put = new Put(row1); - put.add(fam1, qf1, val2); - put.add(fam2, qf1, val3); - put.add(fam2, qf2, val2); - put.add(fam2, qf3, val1); - put.add(fam1, qf3, val1); + put.addColumn(fam1, qf1, val2); + put.addColumn(fam2, qf1, val3); + put.addColumn(fam2, qf2, val2); + put.addColumn(fam2, qf3, val1); + put.addColumn(fam1, qf3, val1); region.put(put); // Multi-column delete @@ -2040,8 +2040,8 @@ public class TestHRegion { byte[] value = Bytes.toBytes("value"); Put put = new Put(row1); - put.add(fam1, qual, 1, value); - put.add(fam1, qual, 2, value); + put.addColumn(fam1, qual, (long) 1, value); + put.addColumn(fam1, qual, (long) 2, value); String method = this.getName(); this.region = initHRegion(tableName, method, CONF, fam1); @@ -2125,15 +2125,15 @@ public class TestHRegion { // add some data: Put put = new Put(row); - put.add(fam, splitA, Bytes.toBytes("reference_A")); + put.addColumn(fam, splitA, Bytes.toBytes("reference_A")); region.put(put); put = new Put(row); - put.add(fam, splitB, Bytes.toBytes("reference_B")); + put.addColumn(fam, splitB, Bytes.toBytes("reference_B")); region.put(put); put = new Put(row); - put.add(fam, serverinfo, Bytes.toBytes("ip_address")); + put.addColumn(fam, serverinfo, Bytes.toBytes("ip_address")); region.put(put); // ok now delete a split: @@ -2156,7 +2156,7 @@ public class TestHRegion { // Assert that after a delete, I can put. put = new Put(row); - put.add(fam, splitA, Bytes.toBytes("reference_A")); + put.addColumn(fam, splitA, Bytes.toBytes("reference_A")); region.put(put); get = new Get(row); result = region.get(get); @@ -2167,7 +2167,7 @@ public class TestHRegion { region.delete(delete); assertEquals(0, region.get(get).size()); - region.put(new Put(row).add(fam, splitA, Bytes.toBytes("reference_A"))); + region.put(new Put(row).addColumn(fam, splitA, Bytes.toBytes("reference_A"))); result = region.get(get); assertEquals(1, result.size()); } finally { @@ -2189,7 +2189,7 @@ public class TestHRegion { // add data in the far future Put put = new Put(row); - put.add(fam, serverinfo, HConstants.LATEST_TIMESTAMP - 5, Bytes.toBytes("value")); + put.addColumn(fam, serverinfo, HConstants.LATEST_TIMESTAMP - 5, Bytes.toBytes("value")); region.put(put); // now delete something in the present @@ -2232,7 +2232,7 @@ public class TestHRegion { // add data with LATEST_TIMESTAMP, put without WAL Put put = new Put(row); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, Bytes.toBytes("value")); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, Bytes.toBytes("value")); region.put(put); // Make sure it shows up with an actual timestamp @@ -2248,7 +2248,7 @@ public class TestHRegion { // code paths, so check both) row = Bytes.toBytes("row2"); put = new Put(row); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, Bytes.toBytes("value")); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, Bytes.toBytes("value")); region.put(put); // Make sure it shows up with an actual timestamp @@ -2284,10 +2284,9 @@ public class TestHRegion { try { try { // no TS specified == use latest. should not error - region.put(new Put(row).add(fam, Bytes.toBytes("qual"), Bytes.toBytes("value"))); + region.put(new Put(row).addColumn(fam, Bytes.toBytes("qual"), Bytes.toBytes("value"))); // TS out of range. should error - region.put(new Put(row).add(fam, Bytes.toBytes("qual"), System.currentTimeMillis() + 2000, - Bytes.toBytes("value"))); + region.put(new Put(row).addColumn(fam, Bytes.toBytes("qual"), System.currentTimeMillis() + 2000, Bytes.toBytes("value"))); fail("Expected IOE for TS out of configured timerange"); } catch (FailedSanityCheckException ioe) { LOG.debug("Received expected exception", ioe); @@ -2318,12 +2317,12 @@ public class TestHRegion { // now create data. Put put = new Put(rowA); - put.add(fam2, null, value); + put.addColumn(fam2, null, value); region.put(put); put = new Put(rowB); - put.add(fam1, null, value); - put.add(fam2, null, value); + put.addColumn(fam1, null, value); + put.addColumn(fam2, null, value); region.put(put); Scan scan = new Scan(); @@ -2362,7 +2361,7 @@ public class TestHRegion { try { EnvironmentEdgeManagerTestHelper.injectEdge(new IncrementingEnvironmentEdge()); Put put = new Put(row); - put.add(fam1, qual1, value1); + put.addColumn(fam1, qual1, value1); region.put(put); // now delete the value: @@ -2370,7 +2369,7 @@ public class TestHRegion { // ok put data: put = new Put(row); - put.add(fam1, qual1, value2); + put.addColumn(fam1, qual1, value2); region.put(put); // ok get: @@ -2488,11 +2487,11 @@ public class TestHRegion { try { // Add to memstore Put put = new Put(row1); - put.add(fam1, col1, null); - put.add(fam1, col2, null); - put.add(fam1, col3, null); - put.add(fam1, col4, null); - put.add(fam1, col5, null); + put.addColumn(fam1, col1, null); + put.addColumn(fam1, col2, null); + put.addColumn(fam1, col3, null); + put.addColumn(fam1, col4, null); + put.addColumn(fam1, col5, null); region.put(put); Get get = new Get(row1); @@ -2703,10 +2702,10 @@ public class TestHRegion { // Putting data in Region Put put = new Put(row1); - put.add(fam1, null, null); - put.add(fam2, null, null); - put.add(fam3, null, null); - put.add(fam4, null, null); + put.addColumn(fam1, null, null); + put.addColumn(fam2, null, null); + put.addColumn(fam3, null, null); + put.addColumn(fam4, null, null); region.put(put); Scan scan = null; @@ -2786,17 +2785,17 @@ public class TestHRegion { // Putting data in Region Put put = null; put = new Put(row1); - put.add(fam1, (byte[]) null, ts, null); - put.add(fam2, (byte[]) null, ts, null); - put.add(fam3, (byte[]) null, ts, null); - put.add(fam4, (byte[]) null, ts, null); + put.addColumn(fam1, (byte[]) null, ts, null); + put.addColumn(fam2, (byte[]) null, ts, null); + put.addColumn(fam3, (byte[]) null, ts, null); + put.addColumn(fam4, (byte[]) null, ts, null); region.put(put); put = new Put(row2); - put.add(fam1, (byte[]) null, ts, null); - put.add(fam2, (byte[]) null, ts, null); - put.add(fam3, (byte[]) null, ts, null); - put.add(fam4, (byte[]) null, ts, null); + put.addColumn(fam1, (byte[]) null, ts, null); + put.addColumn(fam2, (byte[]) null, ts, null); + put.addColumn(fam3, (byte[]) null, ts, null); + put.addColumn(fam4, (byte[]) null, ts, null); region.put(put); Scan scan = new Scan(); @@ -3174,23 +3173,23 @@ public class TestHRegion { byte[] col2 = Bytes.toBytes("Pub222"); Put put = new Put(row1); - put.add(family, col1, Bytes.toBytes(10L)); + put.addColumn(family, col1, Bytes.toBytes(10L)); region.put(put); put = new Put(row2); - put.add(family, col1, Bytes.toBytes(15L)); + put.addColumn(family, col1, Bytes.toBytes(15L)); region.put(put); put = new Put(row3); - put.add(family, col2, Bytes.toBytes(20L)); + put.addColumn(family, col2, Bytes.toBytes(20L)); region.put(put); put = new Put(row4); - put.add(family, col2, Bytes.toBytes(30L)); + put.addColumn(family, col2, Bytes.toBytes(30L)); region.put(put); put = new Put(row5); - put.add(family, col1, Bytes.toBytes(40L)); + put.addColumn(family, col1, Bytes.toBytes(40L)); region.put(put); Scan scan = new Scan(row3, row4); @@ -3309,19 +3308,19 @@ public class TestHRegion { byte[] filtered_val = Bytes.toBytes(3); Put put = new Put(row1); - put.add(cf_essential, col_normal, Bytes.toBytes(1)); - put.add(cf_joined, col_alpha, Bytes.toBytes(1)); + put.addColumn(cf_essential, col_normal, Bytes.toBytes(1)); + put.addColumn(cf_joined, col_alpha, Bytes.toBytes(1)); region.put(put); put = new Put(row2); - put.add(cf_essential, col_alpha, Bytes.toBytes(2)); - put.add(cf_joined, col_normal, Bytes.toBytes(2)); - put.add(cf_alpha, col_alpha, Bytes.toBytes(2)); + put.addColumn(cf_essential, col_alpha, Bytes.toBytes(2)); + put.addColumn(cf_joined, col_normal, Bytes.toBytes(2)); + put.addColumn(cf_alpha, col_alpha, Bytes.toBytes(2)); region.put(put); put = new Put(row3); - put.add(cf_essential, col_normal, filtered_val); - put.add(cf_joined, col_normal, filtered_val); + put.addColumn(cf_essential, col_normal, filtered_val); + put.addColumn(cf_joined, col_normal, filtered_val); region.put(put); // Check two things: @@ -3374,11 +3373,11 @@ public class TestHRegion { for (int i = 0; i < 10; i++) { put = new Put(Bytes.toBytes("r" + Integer.toString(i))); - put.add(cf_first, col_a, Bytes.toBytes(i)); + put.addColumn(cf_first, col_a, Bytes.toBytes(i)); if (i < 5) { - put.add(cf_first, col_b, Bytes.toBytes(i)); - put.add(cf_second, col_a, Bytes.toBytes(i)); - put.add(cf_second, col_b, Bytes.toBytes(i)); + put.addColumn(cf_first, col_b, Bytes.toBytes(i)); + put.addColumn(cf_second, col_a, Bytes.toBytes(i)); + put.addColumn(cf_second, col_b, Bytes.toBytes(i)); } region.put(put); } @@ -3712,7 +3711,7 @@ public class TestHRegion { for (long i = 0; i < numRows; i++) { Put put = new Put(Bytes.toBytes(i)); put.setDurability(Durability.SKIP_WAL); - put.add(family, qual1, Bytes.toBytes(i % 10)); + put.addColumn(family, qual1, Bytes.toBytes(i % 10)); region.put(put); if (i != 0 && i % compactInterval == 0) { @@ -3946,7 +3945,7 @@ public class TestHRegion { byte[] value = Bytes.toBytes(String.valueOf(numPutsFinished)); for (byte[] family : families) { for (byte[] qualifier : qualifiers) { - put.add(family, qualifier, (long) numPutsFinished, value); + put.addColumn(family, qualifier, (long) numPutsFinished, value); } } region.put(put); @@ -4123,7 +4122,7 @@ public class TestHRegion { this.region = initHRegion(tableName, method, CONF, family); try { Put put = new Put(Bytes.toBytes(1L)); - put.add(family, qual1, 1L, Bytes.toBytes(1L)); + put.addColumn(family, qual1, 1L, Bytes.toBytes(1L)); region.put(put); region.flush(true); @@ -4132,7 +4131,7 @@ public class TestHRegion { region.delete(delete); put = new Put(Bytes.toBytes(2L)); - put.add(family, qual1, 2L, Bytes.toBytes(2L)); + put.addColumn(family, qual1, 2L, Bytes.toBytes(2L)); region.put(put); Scan idxScan = new Scan(); @@ -4180,7 +4179,8 @@ public class TestHRegion { for (int j = 0; j < num_unique_rows; j++) { Put put = new Put(Bytes.toBytes("row" + j)); put.setDurability(Durability.SKIP_WAL); - put.add(fam1, qf1, version++, val1); + long ts = version++; + put.addColumn(fam1, qf1, ts, val1); region.put(put); } } @@ -4233,7 +4233,7 @@ public class TestHRegion { Put put = new Put(row); put.setDurability(Durability.SKIP_WAL); for (long idx = 1; idx <= 4; idx++) { - put.add(FAMILY, column, idx, Bytes.toBytes("value-version-" + idx)); + put.addColumn(FAMILY, column, idx, Bytes.toBytes("value-version-" + idx)); } region.put(put); @@ -4280,7 +4280,7 @@ public class TestHRegion { byte col[] = Bytes.toBytes("col1"); Put put = new Put(row); - put.add(familyName, col, 1, Bytes.toBytes("SomeRandomValue")); + put.addColumn(familyName, col, (long) 1, Bytes.toBytes("SomeRandomValue")); region.put(put); region.flush(true); @@ -4327,8 +4327,8 @@ public class TestHRegion { byte col[] = Bytes.toBytes("col1"); Put put = new Put(row); - put.add(fam1, col, 1, Bytes.toBytes("test1")); - put.add(fam2, col, 1, Bytes.toBytes("test2")); + put.addColumn(fam1, col, (long) 1, Bytes.toBytes("test1")); + put.addColumn(fam2, col, (long) 1, Bytes.toBytes("test2")); ht.put(put); HRegion firstRegion = htu.getHBaseCluster().getRegions(TableName.valueOf(this.getName())) @@ -4657,7 +4657,7 @@ public class TestHRegion { put = new Put(row); value = Bytes.toBytes("value0"); - put.add(family, qualifier, 1234567l, value); + put.addColumn(family, qualifier, 1234567l, value); region.put(put); get = new Get(row); get.addColumn(family, qualifier); @@ -4678,7 +4678,7 @@ public class TestHRegion { put = new Put(row); value = Bytes.toBytes("value1"); - put.add(family, qualifier, 1234567l, value); + put.addColumn(family, qualifier, 1234567l, value); region.put(put); get = new Get(row); get.addColumn(family, qualifier); @@ -4766,7 +4766,7 @@ public class TestHRegion { new byte[][] { family }); Put put = new Put(Bytes.toBytes("r1")); - put.add(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); + put.addColumn(family, Bytes.toBytes("q1"), Bytes.toBytes("v1")); put.setDurability(mutationDurability); region.put(put); @@ -4982,7 +4982,7 @@ public class TestHRegion { Put put = new Put(Bytes.toBytes("" + i)); put.setDurability(durability); for (byte[] family : families) { - put.add(family, qf, null); + put.addColumn(family, qf, null); } region.put(put); } @@ -5076,7 +5076,7 @@ public class TestHRegion { // Flush enough files to get up to the threshold, doesn't need compactions for (int i = 0; i < 2; i++) { - Put put = new Put(tableName.toBytes()).add(family, family, tableName.toBytes()); + Put put = new Put(tableName.toBytes()).addColumn(family, family, tableName.toBytes()); region.put(put); fr = region.flush(true); assertTrue(fr.isFlushSucceeded()); @@ -5085,7 +5085,7 @@ public class TestHRegion { // Two flushes after the threshold, compactions are needed for (int i = 0; i < 2; i++) { - Put put = new Put(tableName.toBytes()).add(family, family, tableName.toBytes()); + Put put = new Put(tableName.toBytes()).addColumn(family, family, tableName.toBytes()); region.put(put); fr = region.flush(true); assertTrue(fr.isFlushSucceeded()); @@ -5788,11 +5788,11 @@ public class TestHRegion { try { // setup with one storefile and one memstore, to create scanner and get an earlier readPt Put put = new Put(Bytes.toBytes("19998")); - put.add(cf1, col, Bytes.toBytes("val")); + put.addColumn(cf1, col, Bytes.toBytes("val")); region.put(put); region.flushcache(true, true); Put put2 = new Put(Bytes.toBytes("19997")); - put2.add(cf1, col, Bytes.toBytes("val")); + put2.addColumn(cf1, col, Bytes.toBytes("val")); region.put(put2); Scan scan = new Scan(Bytes.toBytes("19998")); @@ -5803,7 +5803,7 @@ public class TestHRegion { // to check StoreFileScanner.seekToPreviousRow for (int i = 10000; i < 20000; i++) { Put p = new Put(Bytes.toBytes(""+i)); - p.add(cf1, col, Bytes.toBytes(""+i)); + p.addColumn(cf1, col, Bytes.toBytes("" + i)); region.put(p); } region.flushcache(true, true); @@ -5812,7 +5812,7 @@ public class TestHRegion { // to check MemStoreScanner.seekToPreviousRow for (int i = 10000; i < 20000; i++) { Put p = new Put(Bytes.toBytes(""+i)); - p.add(cf1, col, Bytes.toBytes(""+i)); + p.addColumn(cf1, col, Bytes.toBytes("" + i)); region.put(p); } @@ -5939,7 +5939,7 @@ public class TestHRegion { Assert.assertEquals(0L, region.getWriteRequestsCount()); Put put = new Put(row); - put.add(fam, fam, fam); + put.addColumn(fam, fam, fam); Assert.assertEquals(0L, region.getWriteRequestsCount()); region.put(put); @@ -5976,7 +5976,7 @@ public class TestHRegion { assertNotNull(region); // create a file in fam1 for the region before opening in OpenRegionHandler - region.put(new Put(Bytes.toBytes("a")).add(fam1, fam1, fam1)); + region.put(new Put(Bytes.toBytes("a")).addColumn(fam1, fam1, fam1)); region.flush(true); HBaseTestingUtility.closeRegionAndWAL(region); @@ -6081,7 +6081,7 @@ public class TestHRegion { assertNotNull(region); // create a file in fam1 for the region before opening in OpenRegionHandler - region.put(new Put(Bytes.toBytes("a")).add(fam1, fam1, fam1)); + region.put(new Put(Bytes.toBytes("a")).addColumn(fam1, fam1, fam1)); region.flush(true); HBaseTestingUtility.closeRegionAndWAL(region); @@ -6314,14 +6314,14 @@ public class TestHRegion { // TTL tags specify ts in milliseconds new Tag(TagType.TTL_TAG_TYPE, Bytes.toBytes(5000L)) } ))); // Add a cell that will expire after 10 seconds via family setting - region.put(new Put(row).add(fam1, q2, now, HConstants.EMPTY_BYTE_ARRAY)); + region.put(new Put(row).addColumn(fam1, q2, now, HConstants.EMPTY_BYTE_ARRAY)); // Add a cell that will expire in 15 seconds via cell TTL region.put(new Put(row).add(new KeyValue(row, fam1, q3, now + 10000 - 1, HConstants.EMPTY_BYTE_ARRAY, new Tag[] { // TTL tags specify ts in milliseconds new Tag(TagType.TTL_TAG_TYPE, Bytes.toBytes(5000L)) } ))); // Add a cell that will expire in 20 seconds via family setting - region.put(new Put(row).add(fam1, q4, now + 10000 - 1, HConstants.EMPTY_BYTE_ARRAY)); + region.put(new Put(row).addColumn(fam1, q4, now + 10000 - 1, HConstants.EMPTY_BYTE_ARRAY)); // Flush so we are sure store scanning gets this right region.flush(true); @@ -6372,7 +6372,7 @@ public class TestHRegion { // Fun with disappearing increments // Start at 1 - region.put(new Put(row).add(fam1, q1, Bytes.toBytes(1L))); + region.put(new Put(row).addColumn(fam1, q1, Bytes.toBytes(1L))); r = region.get(new Get(row)); byte[] val = r.getValue(fam1, q1); assertNotNull(val); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java index fa152ca..98d98aa 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java @@ -30,7 +30,6 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Result; @@ -146,7 +145,7 @@ public class TestHRegionOnCluster { String value, int verifyNum) throws IOException { System.out.println("=========Putting data :" + row); Put put = new Put(Bytes.toBytes(row)); - put.add(family, Bytes.toBytes("q1"), Bytes.toBytes(value)); + put.addColumn(family, Bytes.toBytes("q1"), Bytes.toBytes(value)); table.put(put); ResultScanner resultScanner = table.getScanner(new Scan()); List results = new ArrayList(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java index 04e9b56..0c2e01c 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionReplayEvents.java @@ -1078,7 +1078,7 @@ public class TestHRegionReplayEvents { long readPoint = region.getMVCC().getReadPoint(); long origSeqId = readPoint + 100; - Put put = new Put(row).add(family, row, row); + Put put = new Put(row).addColumn(family, row, row); put.setDurability(Durability.SKIP_WAL); // we replay with skip wal replay(region, put, origSeqId); @@ -1091,7 +1091,7 @@ public class TestHRegionReplayEvents { // replay an entry that is smaller than current read point // caution: adding an entry below current read point might cause partial dirty reads. Normal // replay does not allow reads while replay is going on. - put = new Put(row2).add(family, row2, row2); + put = new Put(row2).addColumn(family, row2, row2); put.setDurability(Durability.SKIP_WAL); replay(region, put, origSeqId - 50); @@ -1628,7 +1628,7 @@ public class TestHRegionReplayEvents { Put put = new Put(Bytes.toBytes("" + i)); put.setDurability(Durability.SKIP_WAL); for (byte[] family : families) { - put.add(family, qf, EnvironmentEdgeManager.currentTime(), null); + put.addColumn(family, qf, EnvironmentEdgeManager.currentTime(), null); } replay(region, put, i+1); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java index 1fe57d0..67c5f51 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java @@ -36,7 +36,6 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -114,11 +113,11 @@ public class TestJoinedScanners { for (long i = 0; i < rows_to_insert; i++) { Put put = new Put(Bytes.toBytes(Long.toString (i))); if (rand.nextInt(100) <= selectionRatio) { - put.add(cf_essential, col_name, flag_yes); + put.addColumn(cf_essential, col_name, flag_yes); } else { - put.add(cf_essential, col_name, flag_no); + put.addColumn(cf_essential, col_name, flag_no); } - put.add(cf_joined, col_name, val_large); + put.addColumn(cf_joined, col_name, val_large); puts.add(put); if (puts.size() >= insert_batch) { ht.put(puts); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java index 767eb51..3e32772 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestKeepDeletes.java @@ -103,16 +103,16 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); p = new Put(T1, ts+1); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); p = new Put(T1, ts+2); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); p = new Put(T1, ts+4); - p.add(c0, c0, T4); + p.addColumn(c0, c0, T4); region.put(p); // now place a delete marker at ts+2 @@ -156,16 +156,16 @@ public class TestKeepDeletes { // two more puts, this will expire the older puts. p = new Put(T1, ts+5); - p.add(c0, c0, T5); + p.addColumn(c0, c0, T5); region.put(p); p = new Put(T1, ts+6); - p.add(c0, c0, T6); + p.addColumn(c0, c0, T6); region.put(p); // also add an old put again // (which is past the max versions) p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); r = region.get(g); assertTrue(r.isEmpty()); @@ -200,7 +200,7 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); Delete d = new Delete(T1, ts); @@ -245,7 +245,7 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); Get gOne = new Get(T1); @@ -322,13 +322,13 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); p = new Put(T1, ts+2); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); p = new Put(T1, ts+4); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); Delete d = new Delete(T1, ts+1); @@ -456,12 +456,12 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // a put into another store (CF) should have no effect p = new Put(T1, ts-10); - p.add(c1, c0, T1); + p.addColumn(c1, c0, T1); region.put(p); // all the following deletes affect the put @@ -491,7 +491,7 @@ public class TestKeepDeletes { // another put will push out the earlier put... p = new Put(T1, ts+3); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); region.flush(true); @@ -519,12 +519,12 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // a put another (older) row in the same store p = new Put(T2, ts-10); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // all the following deletes affect the put @@ -554,7 +554,7 @@ public class TestKeepDeletes { // another put will push out the earlier put... p = new Put(T1, ts+3); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); region.flush(true); @@ -569,7 +569,7 @@ public class TestKeepDeletes { // another put will push out the earlier put... p = new Put(T1, ts+4); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // this pushed out the column and version marker @@ -596,31 +596,31 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); - p.add(c0, c1, T1); - p.add(c1, c0, T1); - p.add(c1, c1, T1); + p.addColumn(c0, c0, T1); + p.addColumn(c0, c1, T1); + p.addColumn(c1, c0, T1); + p.addColumn(c1, c1, T1); region.put(p); p = new Put(T2, ts); - p.add(c0, c0, T1); - p.add(c0, c1, T1); - p.add(c1, c0, T1); - p.add(c1, c1, T1); + p.addColumn(c0, c0, T1); + p.addColumn(c0, c1, T1); + p.addColumn(c1, c0, T1); + p.addColumn(c1, c1, T1); region.put(p); p = new Put(T1, ts+1); - p.add(c0, c0, T2); - p.add(c0, c1, T2); - p.add(c1, c0, T2); - p.add(c1, c1, T2); + p.addColumn(c0, c0, T2); + p.addColumn(c0, c1, T2); + p.addColumn(c1, c0, T2); + p.addColumn(c1, c1, T2); region.put(p); p = new Put(T2, ts+1); - p.add(c0, c0, T2); - p.add(c0, c1, T2); - p.add(c1, c0, T2); - p.add(c1, c1, T2); + p.addColumn(c0, c0, T2); + p.addColumn(c0, c1, T2); + p.addColumn(c1, c0, T2); + p.addColumn(c1, c1, T2); region.put(p); Delete d = new Delete(T1, ts+2); @@ -678,13 +678,13 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // this prevents marker collection based on earliestPut // (cannot keep earliest put per column in the store file) p = new Put(T1, ts-10); - p.add(c0, c1, T1); + p.addColumn(c0, c1, T1); region.put(p); Delete d = new Delete(T1, ts); @@ -709,14 +709,14 @@ public class TestKeepDeletes { // the 2nd put (and all delete markers following) // will be removed. p = new Put(T1, ts+2); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); // delete, put, delete, delete, put assertEquals(3, countDeleteMarkers(region)); p = new Put(T1, ts+3); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); // This is potentially questionable behavior. @@ -746,7 +746,7 @@ public class TestKeepDeletes { // add one more put p = new Put(T1, ts+4); - p.add(c0, c0, T4); + p.addColumn(c0, c0, T4); region.put(p); region.flush(true); @@ -771,17 +771,17 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime(); Put p = new Put(T1, ts); - p.add(c0, c0, T1); - p.add(c0, c1, T1); - p.add(c1, c0, T1); - p.add(c1, c1, T1); + p.addColumn(c0, c0, T1); + p.addColumn(c0, c1, T1); + p.addColumn(c1, c0, T1); + p.addColumn(c1, c1, T1); region.put(p); p = new Put(T2, ts+1); - p.add(c0, c0, T2); - p.add(c0, c1, T2); - p.add(c1, c0, T2); - p.add(c1, c1, T2); + p.addColumn(c0, c0, T2); + p.addColumn(c0, c1, T2); + p.addColumn(c1, c0, T2); + p.addColumn(c1, c1, T2); region.put(p); // family markers are each family @@ -823,16 +823,16 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime() - 2000; // 2s in the past Put p = new Put(T1, ts); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); p = new Put(T1, ts-1); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); p = new Put(T1, ts-3); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); p = new Put(T1, ts-4); - p.add(c0, c0, T0); + p.addColumn(c0, c0, T0); region.put(p); // all puts now are just retained because of min versions = 3 @@ -861,7 +861,7 @@ public class TestKeepDeletes { r = region.get(g); checkResult(r, c0, c0, T1); p = new Put(T1, ts+1); - p.add(c0, c0, T4); + p.addColumn(c0, c0, T4); region.put(p); region.flush(true); @@ -873,7 +873,7 @@ public class TestKeepDeletes { // this will push out the last put before // family delete marker p = new Put(T1, ts+2); - p.add(c0, c0, T5); + p.addColumn(c0, c0, T5); region.put(p); region.flush(true); @@ -902,12 +902,12 @@ public class TestKeepDeletes { long ts = EnvironmentEdgeManager.currentTime() - 2000; // 2s in the past Put p = new Put(T1, ts); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); // place an old row, to make the family marker expires anyway p = new Put(T2, ts-10); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); checkGet(region, T1, c0, c0, ts+1, T3); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java index cd84eac..661583e 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMinVersions.java @@ -75,15 +75,15 @@ public class TestMinVersions { long ts = EnvironmentEdgeManager.currentTime() - 2000; Put p = new Put(T1, ts); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); p = new Put(T1, ts+1); - p.add(c0, c0, T4); + p.addColumn(c0, c0, T4); region.put(p); p = new Put(T3, ts); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); // now make sure that getClosestBefore(...) get can @@ -124,11 +124,11 @@ public class TestMinVersions { try { Put p = new Put(T1, ts-1); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); p = new Put(T1, ts-3); - p.add(c0, c0, T0); + p.addColumn(c0, c0, T0); region.put(p); // now flush/compact @@ -136,15 +136,15 @@ public class TestMinVersions { region.compact(true); p = new Put(T1, ts); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); p = new Put(T1, ts-2); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); p = new Put(T1, ts-3); - p.add(c0, c0, T0); + p.addColumn(c0, c0, T0); region.put(p); // newest version in the memstore @@ -180,15 +180,15 @@ public class TestMinVersions { try { Put p = new Put(T1, ts-2); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); p = new Put(T1, ts-1); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); p = new Put(T1, ts); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); Delete d = new Delete(T1, ts-1); @@ -240,17 +240,17 @@ public class TestMinVersions { try { // 2nd version Put p = new Put(T1, ts-2); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); // 3rd version p = new Put(T1, ts-1); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); // 4th version p = new Put(T1, ts); - p.add(c0, c0, T4); + p.addColumn(c0, c0, T4); region.put(p); // now flush/compact @@ -259,7 +259,7 @@ public class TestMinVersions { // now put the first version (backdated) p = new Put(T1, ts-3); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // now the latest change is in the memstore, @@ -280,7 +280,7 @@ public class TestMinVersions { checkResult(r, c0, T4,T3); p = new Put(T1, ts+1); - p.add(c0, c0, T5); + p.addColumn(c0, c0, T5); region.put(p); // now the latest version is in the memstore @@ -316,22 +316,22 @@ public class TestMinVersions { // 1st version Put p = new Put(T1, ts-3); - p.add(c0, c0, T1); + p.addColumn(c0, c0, T1); region.put(p); // 2nd version p = new Put(T1, ts-2); - p.add(c0, c0, T2); + p.addColumn(c0, c0, T2); region.put(p); // 3rd version p = new Put(T1, ts-1); - p.add(c0, c0, T3); + p.addColumn(c0, c0, T3); region.put(p); // 4th version p = new Put(T1, ts); - p.add(c0, c0, T4); + p.addColumn(c0, c0, T4); region.put(p); Result r = region.get(new Get(T1)); @@ -408,23 +408,23 @@ public class TestMinVersions { try { Put p = new Put(T1, ts-3); - p.add(c0, c0, T0); - p.add(c1, c1, T0); + p.addColumn(c0, c0, T0); + p.addColumn(c1, c1, T0); region.put(p); p = new Put(T1, ts-2); - p.add(c0, c0, T1); - p.add(c1, c1, T1); + p.addColumn(c0, c0, T1); + p.addColumn(c1, c1, T1); region.put(p); p = new Put(T1, ts-1); - p.add(c0, c0, T2); - p.add(c1, c1, T2); + p.addColumn(c0, c0, T2); + p.addColumn(c1, c1, T2); region.put(p); p = new Put(T1, ts); - p.add(c0, c0, T3); - p.add(c1, c1, T3); + p.addColumn(c0, c0, T3); + p.addColumn(c1, c1, T3); region.put(p); List tss = new ArrayList(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java index d9453b0..4dc233f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestParallelPut.java @@ -115,7 +115,7 @@ public class TestParallelPut { long value = 1L; Put put = new Put(row); - put.add(fam1, qual1, Bytes.toBytes(value)); + put.addColumn(fam1, qual1, Bytes.toBytes(value)); region.put(put); assertGet(this.region, row, fam1, qual1, Bytes.toBytes(value)); @@ -215,7 +215,7 @@ public class TestParallelPut { // put the randombytes and verify that we can read it. This is one // way of ensuring that rwcc manipulation in HRegion.put() is fine. Put put = new Put(rowkey); - put.add(fam1, qual1, value); + put.addColumn(fam1, qual1, value); in[0] = put; try { OperationStatus[] ret = region.batchMutate(in); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java index f238770..bcb8733 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransaction.java @@ -460,7 +460,7 @@ public class TestRegionMergeTransaction { continue; } Put put = new Put(k); - put.add(f, null, k); + put.addColumn(f, null, k); if (r.getWAL() == null) put.setDurability(Durability.SKIP_WAL); r.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java index 2a949a1..ae3ad1a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java @@ -456,7 +456,7 @@ public class TestRegionMergeTransactionOnCluster { private void loadData(Table table) throws IOException { for (int i = 0; i < ROWSIZE; i++) { Put put = new Put(ROWS[i]); - put.add(FAMILYNAME, QUALIFIER, Bytes.toBytes(i)); + put.addColumn(FAMILYNAME, QUALIFIER, Bytes.toBytes(i)); table.put(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java index 4a931a7..870e963 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java @@ -40,7 +40,6 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TestMetaTableAccessor; import org.apache.hadoop.hbase.client.Consistency; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Table; @@ -337,7 +336,7 @@ public class TestRegionReplicas { while (running.get()) { byte[] data = Bytes.toBytes(String.valueOf(key)); Put put = new Put(data); - put.add(f, null, data); + put.addColumn(f, null, data); table.put(put); key++; if (key == endKey) key = startKey; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java index 7a9e61f..67fedcd 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java @@ -119,7 +119,7 @@ public class TestRegionServerMetrics { // Do a first put to be sure that the connection is established, meta is there and so on. Table table = connection.getTable(tName); Put p = new Put(row); - p.add(cfName, qualifier, initValue); + p.addColumn(cfName, qualifier, initValue); table.put(p); metricsRegionServer.getRegionServerWrapper().forceRecompute(); @@ -199,7 +199,7 @@ public class TestRegionServerMetrics { Table t = TEST_UTIL.createTable(tableName, cf); Put p = new Put(row); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); p.setDurability(Durability.SKIP_WAL); t.put(p); @@ -227,7 +227,7 @@ public class TestRegionServerMetrics { //Force a hfile. Table t = TEST_UTIL.createTable(tableName, cf); Put p = new Put(row); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); t.put(p); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -251,15 +251,15 @@ public class TestRegionServerMetrics { Table t = TEST_UTIL.createTable(tableName, cf); Put p = new Put(row); - p.add(cf, qualifier, valOne); + p.addColumn(cf, qualifier, valOne); t.put(p); Put pTwo = new Put(row); - pTwo.add(cf, qualifier, valTwo); + pTwo.addColumn(cf, qualifier, valTwo); t.checkAndPut(row, cf, qualifier, valOne, pTwo); Put pThree = new Put(row); - pThree.add(cf, qualifier, valThree); + pThree.addColumn(cf, qualifier, valThree); t.checkAndPut(row, cf, qualifier, valOne, pThree); metricsRegionServer.getRegionServerWrapper().forceRecompute(); @@ -281,7 +281,7 @@ public class TestRegionServerMetrics { Table t = TEST_UTIL.createTable(tableName, cf); Put p = new Put(row); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); t.put(p); for(int count = 0; count< 13; count++) { @@ -308,7 +308,7 @@ public class TestRegionServerMetrics { Table t = TEST_UTIL.createTable(tableName, cf); Put p = new Put(row); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); t.put(p); for(int count = 0; count< 73; count++) { @@ -334,7 +334,7 @@ public class TestRegionServerMetrics { List puts = new ArrayList<>(); for (int insertCount =0; insertCount < 100; insertCount++) { Put p = new Put(Bytes.toBytes("" + insertCount + "row")); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); puts.add(p); } try (Table t = TEST_UTIL.createTable(tableName, cf)) { @@ -384,7 +384,7 @@ public class TestRegionServerMetrics { List puts = new ArrayList<>(); for (int insertCount =0; insertCount < 100; insertCount++) { Put p = new Put(Bytes.toBytes("" + insertCount + "row")); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); puts.add(p); } try (Table t = TEST_UTIL.createTable(tableName, cf)) { @@ -442,7 +442,7 @@ public class TestRegionServerMetrics { t.setAutoFlush(true, true); for (int insertCount = 0; insertCount < numHfiles; insertCount++) { Put p = new Put(Bytes.toBytes(insertCount)); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); t.put(p); admin.flush(tableName); } @@ -471,7 +471,7 @@ public class TestRegionServerMetrics { for (int insertCount = numHfiles; insertCount < 2 * numHfiles - 1; insertCount++) { Put p = new Put(Bytes.toBytes(insertCount)); - p.add(cf, qualifier, val); + p.addColumn(cf, qualifier, val); t.put(p); admin.flush(tableName); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java index e71089d..e8a2134 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerNoMaster.java @@ -28,7 +28,6 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.NotServingRegionException; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Table; @@ -77,7 +76,7 @@ public class TestRegionServerNoMaster { // Create table then get the single region for our new table. table = HTU.createTable(tableName,HConstants.CATALOG_FAMILY); Put p = new Put(row); - p.add(HConstants.CATALOG_FAMILY, row, row); + p.addColumn(HConstants.CATALOG_FAMILY, row, row); table.put(p); try (RegionLocator locator = HTU.getConnection().getRegionLocator(tableName)) { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRowTooBig.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRowTooBig.java index ab12195..4d3a1c3 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRowTooBig.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRowTooBig.java @@ -92,7 +92,8 @@ public class TestRowTooBig { for (int i = 0; i < 5 ; i++) { Put put = new Put(row1); - put.add(fam1, Bytes.toBytes("col_" + i ), new byte[5 * 1024 * 1024]); + byte[] value = new byte[5 * 1024 * 1024]; + put.addColumn(fam1, Bytes.toBytes("col_" + i), value); region.put(put); region.flush(true); } @@ -138,7 +139,8 @@ public class TestRowTooBig { for (int i = 0; i < 10; i++) { Put put = new Put(row1); for (int j = 0; j < 10 * 10000; j++) { - put.add(fam1, Bytes.toBytes("col_" + i + "_" + j), new byte[10]); + byte[] value = new byte[10]; + put.addColumn(fam1, Bytes.toBytes("col_" + i + "_" + j), value); } region.put(put); region.flush(true); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java index c6c9f50..909ae71 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSCVFWithMiniCluster.java @@ -36,7 +36,6 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -91,27 +90,27 @@ public class TestSCVFWithMiniCluster { /* Add a row with 'a:foo' = false */ Put put = new Put(Bytes.toBytes("1")); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_A, QUALIFIER_FOO, Bytes.toBytes("false")); - put.add(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_")); - put.add(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_")); - put.add(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_A, QUALIFIER_FOO, Bytes.toBytes("false")); + put.addColumn(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_")); puts.add(put); /* Add a row with 'a:foo' = true */ put = new Put(Bytes.toBytes("2")); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_A, QUALIFIER_FOO, Bytes.toBytes("true")); - put.add(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_")); - put.add(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_")); - put.add(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_A, QUALIFIER_FOO, Bytes.toBytes("true")); + put.addColumn(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_")); puts.add(put); /* Add a row with 'a:foo' qualifier not set */ put = new Put(Bytes.toBytes("3")); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_")); - put.add(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_")); - put.add(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_A, QUALIFIER_BAR, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_B, QUALIFIER_FOO, Bytes.toBytes("_flag_")); + put.addColumn(FAMILY_B, QUALIFIER_BAR, Bytes.toBytes("_flag_")); puts.add(put); htable.put(puts); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java index cd6ee15..94e2f83 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScanner.java @@ -259,8 +259,7 @@ public class TestScanner { Put put = new Put(ROW_KEY, System.currentTimeMillis()); - put.add(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER, - REGION_INFO.toByteArray()); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER, REGION_INFO.toByteArray()); region.put(put); // What we just committed is in the memstore. Verify that we can get @@ -285,8 +284,7 @@ public class TestScanner { String address = HConstants.LOCALHOST_IP + ":" + HBaseTestingUtility.randomFreePort(); put = new Put(ROW_KEY, System.currentTimeMillis()); - put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, - Bytes.toBytes(address)); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes.toBytes(address)); // put.add(HConstants.COL_STARTCODE, Bytes.toBytes(START_CODE)); @@ -324,8 +322,7 @@ public class TestScanner { put = new Put(ROW_KEY, System.currentTimeMillis()); - put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, - Bytes.toBytes(address)); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes.toBytes(address)); region.put(put); // Validate again diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerRetriableFailure.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerRetriableFailure.java index afeaa1d..e84ed59 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerRetriableFailure.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestScannerRetriableFailure.java @@ -41,10 +41,7 @@ import org.apache.hadoop.hbase.coprocessor.ObserverContext; import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.LargeTests; -import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread; import org.apache.hadoop.hbase.util.Bytes; -import org.apache.hadoop.hbase.util.FSUtils; -import org.apache.hadoop.hbase.util.FSVisitor; import org.apache.hadoop.hbase.util.TestTableName; import org.junit.AfterClass; @@ -138,7 +135,7 @@ public class TestScannerRetriableFailure { byte[] row = Bytes.toBytes(String.format("%09d", i)); Put put = new Put(row); put.setDurability(Durability.SKIP_WAL); - put.add(FAMILY_NAME, null, row); + put.addColumn(FAMILY_NAME, null, row); table.put(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java index 6b45ece..b31be9d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSeekOptimizations.java @@ -309,7 +309,7 @@ public class TestSeekOptimizations { public void put(String qual, long ts) { if (!putTimestamps.contains(ts)) { - put.add(FAMILY_BYTES, Bytes.toBytes(qual), ts, createValue(ts)); + put.addColumn(FAMILY_BYTES, Bytes.toBytes(qual), ts, createValue(ts)); putTimestamps.add(ts); } if (VERBOSE) { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java index a06783a..2549a4f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java @@ -64,7 +64,6 @@ import org.apache.hadoop.hbase.client.Consistency; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; @@ -88,7 +87,6 @@ import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.Repor import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionResponse; import org.apache.hadoop.hbase.regionserver.compactions.CompactionContext; import org.apache.hadoop.hbase.regionserver.compactions.NoLimitCompactionThroughputController; -import org.apache.hadoop.hbase.security.User; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.util.Bytes; @@ -488,7 +486,7 @@ public class TestSplitTransactionOnCluster { String row = "row" + i; Put p = new Put(row.getBytes()); String val = "Val" + i; - p.add("col".getBytes(), "ql".getBytes(), val.getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), val.getBytes()); table.put(p); admin.flush(userTableName.getName()); Delete d = new Delete(row.getBytes()); @@ -502,13 +500,13 @@ public class TestSplitTransactionOnCluster { .getRegionsOfTable(userTableName); HRegionInfo hRegionInfo = regionsOfTable.get(0); Put p = new Put("row6".getBytes()); - p.add("col".getBytes(), "ql".getBytes(), "val".getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), "val".getBytes()); table.put(p); p = new Put("row7".getBytes()); - p.add("col".getBytes(), "ql".getBytes(), "val".getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), "val".getBytes()); table.put(p); p = new Put("row8".getBytes()); - p.add("col".getBytes(), "ql".getBytes(), "val".getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), "val".getBytes()); table.put(p); admin.flush(userTableName.getName()); admin.split(hRegionInfo.getRegionName(), "row7".getBytes()); @@ -755,16 +753,16 @@ public class TestSplitTransactionOnCluster { private void insertData(final TableName tableName, HBaseAdmin admin, Table t) throws IOException, InterruptedException { Put p = new Put(Bytes.toBytes("row1")); - p.add(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("1")); + p.addColumn(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("1")); t.put(p); p = new Put(Bytes.toBytes("row2")); - p.add(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("2")); + p.addColumn(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("2")); t.put(p); p = new Put(Bytes.toBytes("row3")); - p.add(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("3")); + p.addColumn(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("3")); t.put(p); p = new Put(Bytes.toBytes("row4")); - p.add(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("4")); + p.addColumn(Bytes.toBytes("cf"), Bytes.toBytes("q1"), Bytes.toBytes("4")); t.put(p); admin.flush(tableName); } @@ -968,7 +966,7 @@ public class TestSplitTransactionOnCluster { Table hTable = connection.getTable(desc.getTableName()); for(int i = 1; i < 5; i++) { Put p1 = new Put(("r"+i).getBytes()); - p1.add(Bytes.toBytes("f"), "q1".getBytes(), "v".getBytes()); + p1.addColumn(Bytes.toBytes("f"), "q1".getBytes(), "v".getBytes()); hTable.put(p1); } admin.flush(desc.getTableName()); @@ -1057,8 +1055,8 @@ public class TestSplitTransactionOnCluster { HRegion region = regions.get(0); for(int i = 3;i<9;i++) { Put p = new Put(Bytes.toBytes("row"+i)); - p.add(Bytes.toBytes("f"), Bytes.toBytes("q"), Bytes.toBytes("value"+i)); - p.add(Bytes.toBytes("i_f"), Bytes.toBytes("q"), Bytes.toBytes("value"+i)); + p.addColumn(Bytes.toBytes("f"), Bytes.toBytes("q"), Bytes.toBytes("value" + i)); + p.addColumn(Bytes.toBytes("i_f"), Bytes.toBytes("q"), Bytes.toBytes("value" + i)); region.put(p); } region.flush(true); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFileRefresherChore.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFileRefresherChore.java index 18eda70..74b3df9 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFileRefresherChore.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFileRefresherChore.java @@ -121,7 +121,7 @@ public class TestStoreFileRefresherChore { Put put = new Put(Bytes.toBytes("" + i)); put.setDurability(Durability.SKIP_WAL); for (byte[] family : families) { - put.add(family, qf, null); + put.addColumn(family, qf, null); } region.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java index 32e5855..a85e479 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java @@ -120,7 +120,7 @@ public class TestTags { byte[] value = Bytes.toBytes("value"); table = TEST_UTIL.getConnection().getTable(tableName); Put put = new Put(row); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setAttribute("visibility", Bytes.toBytes("myTag")); table.put(put); admin.flush(tableName); @@ -133,7 +133,7 @@ public class TestTags { Put put1 = new Put(row1); byte[] value1 = Bytes.toBytes("1000dfsdf"); - put1.add(fam, qual, HConstants.LATEST_TIMESTAMP, value1); + put1.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value1); // put1.setAttribute("visibility", Bytes.toBytes("myTag3")); table.put(put1); admin.flush(tableName); @@ -141,7 +141,7 @@ public class TestTags { Put put2 = new Put(row2); byte[] value2 = Bytes.toBytes("1000dfsdf"); - put2.add(fam, qual, HConstants.LATEST_TIMESTAMP, value2); + put2.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value2); put2.setAttribute("visibility", Bytes.toBytes("myTag3")); table.put(put2); admin.flush(tableName); @@ -187,7 +187,7 @@ public class TestTags { table = TEST_UTIL.getConnection().getTable(tableName); Put put = new Put(row); byte[] value = Bytes.toBytes("value"); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); table.put(put); admin.flush(tableName); // We are lacking an API for confirming flush request compaction. @@ -199,14 +199,14 @@ public class TestTags { Put put1 = new Put(row1); byte[] value1 = Bytes.toBytes("1000dfsdf"); - put1.add(fam, qual, HConstants.LATEST_TIMESTAMP, value1); + put1.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value1); table.put(put1); admin.flush(tableName); Thread.sleep(1000); Put put2 = new Put(row2); byte[] value2 = Bytes.toBytes("1000dfsdf"); - put2.add(fam, qual, HConstants.LATEST_TIMESTAMP, value2); + put2.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value2); table.put(put2); admin.flush(tableName); Thread.sleep(1000); @@ -277,13 +277,13 @@ public class TestTags { table = TEST_UTIL.getConnection().getTable(tableName); Put put = new Put(row); byte[] value = Bytes.toBytes("value"); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); int bigTagLen = Short.MAX_VALUE - 5; put.setAttribute("visibility", new byte[bigTagLen]); table.put(put); Put put1 = new Put(row1); byte[] value1 = Bytes.toBytes("1000dfsdf"); - put1.add(fam, qual, HConstants.LATEST_TIMESTAMP, value1); + put1.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value1); table.put(put1); admin.flush(tableName); // We are lacking an API for confirming flush request compaction. @@ -295,18 +295,18 @@ public class TestTags { put1 = new Put(row2); value1 = Bytes.toBytes("1000dfsdf"); - put1.add(fam, qual, HConstants.LATEST_TIMESTAMP, value1); + put1.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value1); table.put(put1); admin.flush(tableName); Thread.sleep(1000); Put put2 = new Put(rowd); byte[] value2 = Bytes.toBytes("1000dfsdf"); - put2.add(fam, qual, HConstants.LATEST_TIMESTAMP, value2); + put2.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value2); table.put(put2); put2 = new Put(rowe); value2 = Bytes.toBytes("1000dfsddfdf"); - put2.add(fam, qual, HConstants.LATEST_TIMESTAMP, value2); + put2.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value2); put.setAttribute("visibility", Bytes.toBytes("ram")); table.put(put2); admin.flush(tableName); @@ -390,7 +390,7 @@ public class TestTags { table = TEST_UTIL.getConnection().getTable(tableName); Put put = new Put(row1); byte[] v = Bytes.toBytes(2L); - put.add(f, q, v); + put.addColumn(f, q, v); put.setAttribute("visibility", Bytes.toBytes("tag1")); table.put(put); Increment increment = new Increment(row1); @@ -430,7 +430,7 @@ public class TestTags { put = new Put(row2); v = Bytes.toBytes(2L); - put.add(f, q, v); + put.addColumn(f, q, v); table.put(put); increment = new Increment(row2); increment.add(new KeyValue(row2, f, q, 1234L, v)); @@ -452,7 +452,7 @@ public class TestTags { // Test Append byte[] row3 = Bytes.toBytes("r3"); put = new Put(row3); - put.add(f, q, Bytes.toBytes("a")); + put.addColumn(f, q, Bytes.toBytes("a")); put.setAttribute("visibility", Bytes.toBytes("tag1")); table.put(put); Append append = new Append(row3); @@ -492,7 +492,7 @@ public class TestTags { byte[] row4 = Bytes.toBytes("r4"); put = new Put(row4); - put.add(f, q, Bytes.toBytes("a")); + put.addColumn(f, q, Bytes.toBytes("a")); table.put(put); append = new Append(row4); append.add(new KeyValue(row4, f, q, 1234L, v)); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java index cfa1695..ca7b3b1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java @@ -78,7 +78,8 @@ public class TestWideScanner extends HBaseTestCase { for (j = 0; j < 100; j++) { Put put = new Put(row); put.setDurability(Durability.SKIP_WAL); - put.add(COLUMNS[rng.nextInt(COLUMNS.length)], b, ++ts, b); + long ts1 = ++ts; + put.addColumn(COLUMNS[rng.nextInt(COLUMNS.length)], b, ts1, b); region.put(put); count++; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactionWithThroughputController.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactionWithThroughputController.java index 93c8ebc..9374fdb 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactionWithThroughputController.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactionWithThroughputController.java @@ -35,11 +35,9 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.regionserver.DefaultStoreEngine; -import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.HStore; import org.apache.hadoop.hbase.regionserver.Region; @@ -93,7 +91,7 @@ public class TestCompactionWithThroughputController { for (int j = 0; j < 10; j++) { byte[] value = new byte[128 * 1024]; rand.nextBytes(value); - table.put(new Put(Bytes.toBytes(i * 10 + j)).add(family, qualifier, value)); + table.put(new Put(Bytes.toBytes(i * 10 + j)).addColumn(family, qualifier, value)); } admin.flush(tableName); } @@ -209,18 +207,21 @@ public class TestCompactionWithThroughputController { assertEquals(10L * 1024 * 1024, throughputController.maxThroughput, EPSILON); Table table = conn.getTable(tableName); for (int i = 0; i < 5; i++) { - table.put(new Put(Bytes.toBytes(i)).add(family, qualifier, new byte[0])); + byte[] value = new byte[0]; + table.put(new Put(Bytes.toBytes(i)).addColumn(family, qualifier, value)); TEST_UTIL.flush(tableName); } Thread.sleep(2000); assertEquals(15L * 1024 * 1024, throughputController.maxThroughput, EPSILON); - table.put(new Put(Bytes.toBytes(5)).add(family, qualifier, new byte[0])); + byte[] value1 = new byte[0]; + table.put(new Put(Bytes.toBytes(5)).addColumn(family, qualifier, value1)); TEST_UTIL.flush(tableName); Thread.sleep(2000); assertEquals(20L * 1024 * 1024, throughputController.maxThroughput, EPSILON); - table.put(new Put(Bytes.toBytes(6)).add(family, qualifier, new byte[0])); + byte[] value = new byte[0]; + table.put(new Put(Bytes.toBytes(6)).addColumn(family, qualifier, value)); TEST_UTIL.flush(tableName); Thread.sleep(2000); assertEquals(Double.MAX_VALUE, throughputController.maxThroughput, EPSILON); @@ -261,27 +262,35 @@ public class TestCompactionWithThroughputController { assertEquals(0.0, store.getCompactionPressure(), EPSILON); Table table = conn.getTable(tableName); for (int i = 0; i < 4; i++) { - table.put(new Put(Bytes.toBytes(i)).add(family, qualifier, new byte[0])); - table.put(new Put(Bytes.toBytes(100 + i)).add(family, qualifier, new byte[0])); + byte[] value1 = new byte[0]; + table.put(new Put(Bytes.toBytes(i)).addColumn(family, qualifier, value1)); + byte[] value = new byte[0]; + table.put(new Put(Bytes.toBytes(100 + i)).addColumn(family, qualifier, value)); TEST_UTIL.flush(tableName); } assertEquals(8, store.getStorefilesCount()); assertEquals(0.0, store.getCompactionPressure(), EPSILON); - table.put(new Put(Bytes.toBytes(4)).add(family, qualifier, new byte[0])); - table.put(new Put(Bytes.toBytes(104)).add(family, qualifier, new byte[0])); + byte[] value5 = new byte[0]; + table.put(new Put(Bytes.toBytes(4)).addColumn(family, qualifier, value5)); + byte[] value4 = new byte[0]; + table.put(new Put(Bytes.toBytes(104)).addColumn(family, qualifier, value4)); TEST_UTIL.flush(tableName); assertEquals(10, store.getStorefilesCount()); assertEquals(0.5, store.getCompactionPressure(), EPSILON); - table.put(new Put(Bytes.toBytes(5)).add(family, qualifier, new byte[0])); - table.put(new Put(Bytes.toBytes(105)).add(family, qualifier, new byte[0])); + byte[] value3 = new byte[0]; + table.put(new Put(Bytes.toBytes(5)).addColumn(family, qualifier, value3)); + byte[] value2 = new byte[0]; + table.put(new Put(Bytes.toBytes(105)).addColumn(family, qualifier, value2)); TEST_UTIL.flush(tableName); assertEquals(12, store.getStorefilesCount()); assertEquals(1.0, store.getCompactionPressure(), EPSILON); - table.put(new Put(Bytes.toBytes(6)).add(family, qualifier, new byte[0])); - table.put(new Put(Bytes.toBytes(106)).add(family, qualifier, new byte[0])); + byte[] value1 = new byte[0]; + table.put(new Put(Bytes.toBytes(6)).addColumn(family, qualifier, value1)); + byte[] value = new byte[0]; + table.put(new Put(Bytes.toBytes(106)).addColumn(family, qualifier, value)); TEST_UTIL.flush(tableName); assertEquals(14, store.getStorefilesCount()); assertEquals(2.0, store.getCompactionPressure(), EPSILON); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestDurability.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestDurability.java index 72e4330..0662716 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestDurability.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestDurability.java @@ -218,7 +218,7 @@ public class TestDurability { private Put newPut(Durability durability) { Put p = new Put(ROW); - p.add(FAMILY, COL, COL); + p.addColumn(FAMILY, COL, COL); if (durability != null) { p.setDurability(durability); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java index 8106acf..9dccffe 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java @@ -145,13 +145,13 @@ public class TestLogRollAbort { FSUtils.isAppendSupported(TEST_UTIL.getConfiguration())); Put p = new Put(Bytes.toBytes("row2001")); - p.add(HConstants.CATALOG_FAMILY, Bytes.toBytes("col"), Bytes.toBytes(2001)); + p.addColumn(HConstants.CATALOG_FAMILY, Bytes.toBytes("col"), Bytes.toBytes(2001)); table.put(p); log.sync(); p = new Put(Bytes.toBytes("row2002")); - p.add(HConstants.CATALOG_FAMILY, Bytes.toBytes("col"), Bytes.toBytes(2002)); + p.addColumn(HConstants.CATALOG_FAMILY, Bytes.toBytes("col"), Bytes.toBytes(2002)); table.put(p); dfsCluster.restartDataNodes(); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollPeriod.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollPeriod.java index 5216ee5..f7e5ff7 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollPeriod.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollPeriod.java @@ -27,12 +27,9 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseTestingUtility; -import org.apache.hadoop.hbase.HRegionInfo; -import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.regionserver.HRegionServer; @@ -111,7 +108,7 @@ public class TestLogRollPeriod { long row = 0; while (!interrupted()) { Put p = new Put(Bytes.toBytes(String.format("row%d", row))); - p.add(Bytes.toBytes(family), Bytes.toBytes("col"), Bytes.toBytes(row)); + p.addColumn(Bytes.toBytes(family), Bytes.toBytes("col"), Bytes.toBytes(row)); table.put(p); row++; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java index a5d366b..3ab49c0 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java @@ -268,11 +268,11 @@ public class TestLogRolling { for (int i = 0; i < 10; i++) { Put put = new Put(Bytes.toBytes("row" + String.format("%1$04d", (start + i)))); - put.add(HConstants.CATALOG_FAMILY, null, value); + put.addColumn(HConstants.CATALOG_FAMILY, null, value); table.put(put); } Put tmpPut = new Put(Bytes.toBytes("tmprow")); - tmpPut.add(HConstants.CATALOG_FAMILY, null, value); + tmpPut.addColumn(HConstants.CATALOG_FAMILY, null, value); long startTime = System.currentTimeMillis(); long remaining = timeout; while (remaining > 0) { @@ -631,7 +631,7 @@ public class TestLogRolling { private void doPut(Table table, int i) throws IOException { Put put = new Put(Bytes.toBytes("row" + String.format("%1$04d", i))); - put.add(HConstants.CATALOG_FAMILY, null, value); + put.addColumn(HConstants.CATALOG_FAMILY, null, value); table.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java index 1c97a2d..57139df 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java @@ -177,7 +177,7 @@ public class TestWALReplay { TEST_UTIL.createTable(tableName, familys); Table htable = TEST_UTIL.getConnection().getTable(tableName); Put put = new Put(Bytes.toBytes("r1")); - put.add(family1, qualifier, value); + put.addColumn(family1, qualifier, value); htable.put(put); ResultScanner resultScanner = htable.getScanner(new Scan()); int count = 0; @@ -353,7 +353,7 @@ public class TestWALReplay { // Add an edit so something in the WAL byte [] row = tableName.getName(); - region.put((new Put(row)).add(family, family, family)); + region.put((new Put(row)).addColumn(family, family, family)); wal.sync(); final int rowsInsertedCount = 11; @@ -412,7 +412,7 @@ public class TestWALReplay { // Add an edit so something in the WAL byte [] row = tableName.getName(); byte [] family = htd.getFamilies().iterator().next().getName(); - region.put((new Put(row)).add(family, family, family)); + region.put((new Put(row)).addColumn(family, family, family)); wal.sync(); List > hfs= new ArrayList>(1); @@ -686,8 +686,7 @@ public class TestWALReplay { htd.getFamilies()); for (int i = 0; i < writtenRowCount; i++) { Put put = new Put(Bytes.toBytes(tableName + Integer.toString(i))); - put.add(families.get(i % families.size()).getName(), Bytes.toBytes("q"), - Bytes.toBytes("val")); + put.addColumn(families.get(i % families.size()).getName(), Bytes.toBytes("q"), Bytes.toBytes("val")); region.put(put); } @@ -712,8 +711,7 @@ public class TestWALReplay { int moreRow = 10; for (int i = writtenRowCount; i < writtenRowCount + moreRow; i++) { Put put = new Put(Bytes.toBytes(tableName + Integer.toString(i))); - put.add(families.get(i % families.size()).getName(), Bytes.toBytes("q"), - Bytes.toBytes("val")); + put.addColumn(families.get(i % families.size()).getName(), Bytes.toBytes("q"), Bytes.toBytes("val")); region.put(put); } writtenRowCount += moreRow; @@ -1008,7 +1006,7 @@ public class TestWALReplay { for (int j = 0; j < count; j++) { byte[] qualifier = Bytes.toBytes(qualifierPrefix + Integer.toString(j)); Put p = new Put(rowName); - p.add(family, qualifier, ee.currentTime(), rowName); + p.addColumn(family, qualifier, ee.currentTime(), rowName); r.put(p); puts.add(p); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java index c7ffe25..184fd14 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java @@ -195,7 +195,7 @@ public class TestMasterReplication { // cluster 0 putAndWait(row3, famName, htables[0], htables[1]); // now add a local edit to cluster 1 - htables[1].put(new Put(row4).add(famName, row4, row4)); + htables[1].put(new Put(row4).addColumn(famName, row4, row4)); // re-enable replication from cluster 2 to cluster 0 enablePeer("1", 2); // without HBASE-9158 the edit for row4 would have been marked with @@ -373,7 +373,7 @@ public class TestMasterReplication { private void putAndWait(byte[] row, byte[] fam, Table source, Table target) throws Exception { Put put = new Put(row); - put.add(fam, row, row); + put.addColumn(fam, row, row); source.put(put); wait(row, target, false); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java index b670e65..82a0680 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java @@ -36,8 +36,6 @@ import org.apache.hadoop.hbase.*; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Table; @@ -173,7 +171,7 @@ public class TestMultiSlaveReplication { checkRow(row3,1,htable3); Put p = new Put(row); - p.add(famName, row, row); + p.addColumn(famName, row, row); htable1.put(p); // now roll the logs again rollWALAndWait(utility1, htable1.getName(), row); @@ -301,7 +299,7 @@ public class TestMultiSlaveReplication { private void putAndWait(byte[] row, byte[] fam, Table source, Table... targets) throws Exception { Put put = new Put(row); - put.add(fam, row, row); + put.addColumn(fam, row, row); source.put(put); Get get = new Get(row); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java index 5f42511..29a052b 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java @@ -439,7 +439,7 @@ public class TestPerTableCFReplication { Table source, Table... targets) throws Exception { Put put = new Put(row); - put.add(fam, row, val); + put.addColumn(fam, row, val); source.put(put); Get get = new Get(row); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationChangingPeerRegionservers.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationChangingPeerRegionservers.java index 4bb1842..ba2a7c1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationChangingPeerRegionservers.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationChangingPeerRegionservers.java @@ -115,7 +115,7 @@ public class TestReplicationChangingPeerRegionservers extends TestReplicationBas private void doPutTest(byte[] row) throws IOException, InterruptedException { Put put = new Put(row); - put.add(famName, row, row); + put.addColumn(famName, row, row); if (htable1 == null) { htable1 = utility1.getConnection().getTable(tableName); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationDisableInactivePeer.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationDisableInactivePeer.java index 3378c3f..06e1698 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationDisableInactivePeer.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationDisableInactivePeer.java @@ -54,7 +54,7 @@ public class TestReplicationDisableInactivePeer extends TestReplicationBase { byte[] rowkey = Bytes.toBytes("disable inactive peer"); Put put = new Put(rowkey); - put.add(famName, row, row); + put.addColumn(famName, row, row); htable1.put(put); // wait for the sleep interval of the master cluster to become long diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java index 16564e0..66adf70 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java @@ -34,7 +34,6 @@ import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; -import org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.ReplicationTests; import org.apache.hadoop.hbase.util.Bytes; @@ -218,7 +217,7 @@ public class TestReplicationEndpoint extends TestReplicationBase { private void doPut(final Connection connection, final byte [] row) throws IOException { try (Table t = connection.getTable(tableName)) { Put put = new Put(row); - put.add(famName, row, row); + put.addColumn(famName, row, row); t.put(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java index 46b7956..12470fe 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java @@ -123,15 +123,15 @@ public class TestReplicationSmallTests extends TestReplicationBase { long t = EnvironmentEdgeManager.currentTime(); // create three versions for "row" Put put = new Put(row); - put.add(famName, row, t, v1); + put.addColumn(famName, row, t, v1); htable1.put(put); put = new Put(row); - put.add(famName, row, t+1, v2); + put.addColumn(famName, row, t + 1, v2); htable1.put(put); put = new Put(row); - put.add(famName, row, t+2, v3); + put.addColumn(famName, row, t + 2, v3); htable1.put(put); Get get = new Get(row); @@ -203,7 +203,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { public void testSimplePutDelete() throws Exception { LOG.info("testSimplePutDelete"); Put put = new Put(row); - put.add(famName, row, row); + put.addColumn(famName, row, row); htable1 = utility1.getConnection().getTable(tableName); htable1.put(put); @@ -252,7 +252,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { List puts = new ArrayList<>(); for (int i = 0; i < NB_ROWS_IN_BATCH; i++) { Put put = new Put(Bytes.toBytes(i)); - put.add(famName, row, row); + put.addColumn(famName, row, row); puts.add(put); } htable1.put(puts); @@ -295,7 +295,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { byte[] rowkey = Bytes.toBytes("disable enable"); Put put = new Put(rowkey); - put.add(famName, row, row); + put.addColumn(famName, row, row); htable1.put(put); Get get = new Get(rowkey); @@ -338,7 +338,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { Thread.sleep(SLEEP_TIME); byte[] rowKey = Bytes.toBytes("Won't be replicated"); Put put = new Put(rowKey); - put.add(famName, row, row); + put.addColumn(famName, row, row); htable1.put(put); Get get = new Get(rowKey); @@ -359,7 +359,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { Thread.sleep(SLEEP_TIME); rowKey = Bytes.toBytes("do rep"); put = new Put(rowKey); - put.add(famName, row, row); + put.addColumn(famName, row, row); LOG.info("Adding new row"); htable1.put(put); @@ -391,7 +391,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { List puts = new ArrayList(); for (int i = 0; i < NB_ROWS_IN_BIG_BATCH; i++) { Put put = new Put(Bytes.toBytes(i)); - put.add(famName, row, row); + put.addColumn(famName, row, row); puts.add(put); } htable1.setWriteBufferSize(1024); @@ -472,8 +472,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { for (Result result : rs) { put = new Put(result.getRow()); Cell firstVal = result.rawCells()[0]; - put.add(CellUtil.cloneFamily(firstVal), - CellUtil.cloneQualifier(firstVal), Bytes.toBytes("diff data")); + put.addColumn(CellUtil.cloneFamily(firstVal), CellUtil.cloneQualifier(firstVal), Bytes.toBytes("diff data")); htable2.put(put); } Delete delete = new Delete(put.getRow()); @@ -579,7 +578,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { for (int i = 0; i < NB_ROWS_IN_BATCH; i++) { p = new Put(Bytes.toBytes("row" + i)); - p.add(famName, qualName, Bytes.toBytes("val" + i)); + p.addColumn(famName, qualName, Bytes.toBytes("val" + i)); htable1.put(p); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java index 26acdab..13545b5 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java @@ -225,21 +225,21 @@ public class TestReplicationSyncUpTool extends TestReplicationBase { // 100 + 1 row to t1_syncup for (int i = 0; i < NB_ROWS_IN_BATCH; i++) { p = new Put(Bytes.toBytes("row" + i)); - p.add(famName, qualName, Bytes.toBytes("val" + i)); + p.addColumn(famName, qualName, Bytes.toBytes("val" + i)); ht1Source.put(p); } p = new Put(Bytes.toBytes("row" + 9999)); - p.add(noRepfamName, qualName, Bytes.toBytes("val" + 9999)); + p.addColumn(noRepfamName, qualName, Bytes.toBytes("val" + 9999)); ht1Source.put(p); // 200 + 1 row to t2_syncup for (int i = 0; i < NB_ROWS_IN_BATCH * 2; i++) { p = new Put(Bytes.toBytes("row" + i)); - p.add(famName, qualName, Bytes.toBytes("val" + i)); + p.addColumn(famName, qualName, Bytes.toBytes("val" + i)); ht2Source.put(p); } p = new Put(Bytes.toBytes("row" + 9999)); - p.add(noRepfamName, qualName, Bytes.toBytes("val" + 9999)); + p.addColumn(noRepfamName, qualName, Bytes.toBytes("val" + 9999)); ht2Source.put(p); // ensure replication completed @@ -351,22 +351,22 @@ public class TestReplicationSyncUpTool extends TestReplicationBase { // we should see 100 + 2 rows now for (int i = 0; i < NB_ROWS_IN_BATCH; i++) { p = new Put(Bytes.toBytes("row" + i)); - p.add(famName, qualName, Bytes.toBytes("val" + i)); + p.addColumn(famName, qualName, Bytes.toBytes("val" + i)); ht1Source.put(p); } p = new Put(Bytes.toBytes("row" + 9998)); - p.add(noRepfamName, qualName, Bytes.toBytes("val" + 9998)); + p.addColumn(noRepfamName, qualName, Bytes.toBytes("val" + 9998)); ht1Source.put(p); // another 200 + 1 row to t1_syncup // we should see 200 + 2 rows now for (int i = 0; i < NB_ROWS_IN_BATCH * 2; i++) { p = new Put(Bytes.toBytes("row" + i)); - p.add(famName, qualName, Bytes.toBytes("val" + i)); + p.addColumn(famName, qualName, Bytes.toBytes("val" + i)); ht2Source.put(p); } p = new Put(Bytes.toBytes("row" + 9998)); - p.add(noRepfamName, qualName, Bytes.toBytes("val" + 9998)); + p.addColumn(noRepfamName, qualName, Bytes.toBytes("val" + 9998)); ht2Source.put(p); int rowCount_ht1Source = utility1.countRows(ht1Source); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java index 7f5b59c..988373f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java @@ -167,7 +167,7 @@ public class TestReplicationWithTags { LOG.info("testSimplePutDelete"); Put put = new Put(ROW); put.setAttribute("visibility", Bytes.toBytes("myTag3")); - put.add(FAMILY, ROW, ROW); + put.addColumn(FAMILY, ROW, ROW); htable1 = utility1.getConnection().getTable(TABLE_NAME); htable1.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessControlFilter.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessControlFilter.java index 6b0534e..cf01463 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessControlFilter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessControlFilter.java @@ -115,8 +115,8 @@ public class TestAccessControlFilter extends SecureTestUtil { List puts = new ArrayList(100); for (int i=0; i<100; i++) { Put p = new Put(Bytes.toBytes(i)); - p.add(FAMILY, PRIVATE_COL, Bytes.toBytes("secret "+i)); - p.add(FAMILY, PUBLIC_COL, Bytes.toBytes("info "+i)); + p.addColumn(FAMILY, PRIVATE_COL, Bytes.toBytes("secret " + i)); + p.addColumn(FAMILY, PUBLIC_COL, Bytes.toBytes("info " + i)); puts.add(p); } table.put(puts); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java index a25da37..8ecc6e3 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java @@ -920,7 +920,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put p = new Put(TEST_ROW); - p.add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(1)); + p.addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(1)); try(Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(TEST_TABLE)) { t.put(p); @@ -984,7 +984,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put p = new Put(TEST_ROW); - p.add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(1)); + p.addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes(1)); try(Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(TEST_TABLE);) { t.checkAndPut(TEST_ROW, TEST_FAMILY, TEST_QUALIFIER, @@ -1122,7 +1122,7 @@ public class TestAccessController extends SecureTestUtil { byte[] row = TEST_ROW; byte[] qualifier = TEST_QUALIFIER; Put put = new Put(row); - put.add(TEST_FAMILY, qualifier, Bytes.toBytes(1)); + put.addColumn(TEST_FAMILY, qualifier, Bytes.toBytes(1)); Append append = new Append(row); append.add(TEST_FAMILY, qualifier, Bytes.toBytes(2)); try(Connection conn = ConnectionFactory.createConnection(conf); @@ -1251,8 +1251,8 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put p = new Put(Bytes.toBytes("a")); - p.add(family1, qualifier, Bytes.toBytes("v1")); - p.add(family2, qualifier, Bytes.toBytes("v2")); + p.addColumn(family1, qualifier, Bytes.toBytes("v1")); + p.addColumn(family2, qualifier, Bytes.toBytes("v2")); try (Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(tableName);) { @@ -1266,7 +1266,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put p = new Put(Bytes.toBytes("a")); - p.add(family1, qualifier, Bytes.toBytes("v1")); + p.addColumn(family1, qualifier, Bytes.toBytes("v1")); try (Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(tableName)) { @@ -1280,7 +1280,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put p = new Put(Bytes.toBytes("a")); - p.add(family2, qualifier, Bytes.toBytes("v2")); + p.addColumn(family2, qualifier, Bytes.toBytes("v2")); try (Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(tableName);) { t.put(p); @@ -1515,7 +1515,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put p = new Put(TEST_ROW); - p.add(family1, qualifier, Bytes.toBytes("v1")); + p.addColumn(family1, qualifier, Bytes.toBytes("v1")); try (Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(tableName)) { t.put(p); @@ -2145,7 +2145,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { Put put = new Put(Bytes.toBytes("test")); - put.add(TEST_FAMILY, Bytes.toBytes("qual"), Bytes.toBytes("value")); + put.addColumn(TEST_FAMILY, Bytes.toBytes("qual"), Bytes.toBytes("value")); table.put(put); return null; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java index 7151c46..2299906 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java @@ -288,8 +288,7 @@ public class TestAccessController2 extends SecureTestUtil { try (Connection conn = ConnectionFactory.createConnection(conf); Table t = conn.getTable(AccessControlLists.ACL_TABLE_NAME)) { - t.put(new Put(TEST_ROW).add(AccessControlLists.ACL_LIST_FAMILY, TEST_QUALIFIER, - TEST_VALUE)); + t.put(new Put(TEST_ROW).addColumn(AccessControlLists.ACL_LIST_FAMILY, TEST_QUALIFIER, TEST_VALUE)); return null; } finally { } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java index 69b0b5e..e0aa040 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java @@ -169,20 +169,20 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { Table t = connection.getTable(TEST_TABLE.getTableName())) { Put p; // with ro ACL - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(writePerms); t.put(p); // with ro ACL - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(readPerms); t.put(p); - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(writePerms); t.put(p); - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(readPerms); t.put(p); - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(writePerms); t.put(p); } @@ -228,13 +228,13 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { try(Connection connection = ConnectionFactory.createConnection(conf); Table t = connection.getTable(TEST_TABLE.getTableName())) { Put p; - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(writePerms); t.put(p); - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(readPerms); t.put(p); - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); p.setACL(writePerms); t.put(p); } @@ -275,15 +275,15 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { try (Table t = connection.getTable(TEST_TABLE.getTableName())) { // with rw ACL for "user1" Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, ZERO); p.setACL(user1.getShortName(), new Permission(Permission.Action.READ, Permission.Action.WRITE)); t.put(p); // with rw ACL for "user1" p = new Put(TEST_ROW2); - p.add(TEST_FAMILY1, TEST_Q1, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, ZERO); p.setACL(user1.getShortName(), new Permission(Permission.Action.READ, Permission.Action.WRITE)); t.put(p); @@ -300,8 +300,8 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { try (Table t = connection.getTable(TEST_TABLE.getTableName())) { // with rw ACL for "user1", "user2" and "@group" Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, ZERO); Map perms = prepareCellPermissions(new String[] { user1.getShortName(), user2.getShortName(), AuthUtil.toGroupEntry(GROUP) }, Action.READ, Action.WRITE); @@ -309,8 +309,8 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { t.put(p); // with rw ACL for "user1", "user2" and "@group" p = new Put(TEST_ROW2); - p.add(TEST_FAMILY1, TEST_Q1, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, ZERO); p.setACL(perms); t.put(p); } @@ -391,18 +391,17 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { try (Connection connection = ConnectionFactory.createConnection(conf)) { try (Table t = connection.getTable(TEST_TABLE.getTableName())) { // Store a read write ACL without a timestamp, server will use current time - Put p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q2, ONE); + Put p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q2, ONE); Map readAndWritePerms = prepareCellPermissions(usersAndGroups, Action.READ, Action.WRITE); p.setACL(readAndWritePerms); t.put(p); - p = new Put(TEST_ROW).add(TEST_FAMILY2, TEST_Q2, ONE); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY2, TEST_Q2, ONE); p.setACL(readAndWritePerms); t.put(p); LOG.info("Stored at current time"); // Store read only ACL at a future time - p = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, - EnvironmentEdgeManager.currentTime() + 1000000, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, EnvironmentEdgeManager.currentTime() + 1000000, ZERO); p.setACL(prepareCellPermissions(new String[]{ USER_OTHER.getShortName(), AuthUtil.toGroupEntry(GROUP)}, Action.READ)); t.put(p); @@ -484,8 +483,8 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { try (Table t = connection.getTable(TEST_TABLE.getTableName())) { // This version (TS = 123) with rw ACL for USER_OTHER and USER_OTHER2 Put p = new Put(TEST_ROW); - p.add(TEST_FAMILY1, TEST_Q1, 123L, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, 123L, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, 123L, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, 123L, ZERO); p.setACL(prepareCellPermissions( new String[] { USER_OTHER.getShortName(), AuthUtil.toGroupEntry(GROUP), USER_OTHER2.getShortName() }, Permission.Action.READ, Permission.Action.WRITE)); @@ -493,8 +492,8 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { // This version (TS = 125) with rw ACL for USER_OTHER p = new Put(TEST_ROW); - p.add(TEST_FAMILY1, TEST_Q1, 125L, ONE); - p.add(TEST_FAMILY1, TEST_Q2, 125L, ONE); + p.addColumn(TEST_FAMILY1, TEST_Q1, 125L, ONE); + p.addColumn(TEST_FAMILY1, TEST_Q2, 125L, ONE); p.setACL(prepareCellPermissions( new String[] { USER_OTHER.getShortName(), AuthUtil.toGroupEntry(GROUP) }, Action.READ, Action.WRITE)); @@ -502,8 +501,8 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { // This version (TS = 127) with rw ACL for USER_OTHER p = new Put(TEST_ROW); - p.add(TEST_FAMILY1, TEST_Q1, 127L, TWO); - p.add(TEST_FAMILY1, TEST_Q2, 127L, TWO); + p.addColumn(TEST_FAMILY1, TEST_Q1, 127L, TWO); + p.addColumn(TEST_FAMILY1, TEST_Q2, 127L, TWO); p.setACL(prepareCellPermissions( new String[] { USER_OTHER.getShortName(), AuthUtil.toGroupEntry(GROUP) }, Action.READ, Action.WRITE)); @@ -570,36 +569,36 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { new String[] { user2.getShortName(), AuthUtil.toGroupEntry(GROUP), USER_OWNER.getShortName() }, Action.READ, Action.WRITE); Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 123, ZERO); p.setACL(permsU1andOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 123, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY2, TEST_Q1, 123, ZERO); - p.add(TEST_FAMILY2, TEST_Q2, 123, ZERO); + p.addColumn(TEST_FAMILY2, TEST_Q1, (long) 123, ZERO); + p.addColumn(TEST_FAMILY2, TEST_Q2, (long) 123, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY2, TEST_Q1, 125, ZERO); - p.add(TEST_FAMILY2, TEST_Q2, 125, ZERO); + p.addColumn(TEST_FAMILY2, TEST_Q1, (long) 125, ZERO); + p.addColumn(TEST_FAMILY2, TEST_Q2, (long) 125, ZERO); p.setACL(permsU1andOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 127, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 127, ZERO); p.setACL(permsU1andOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY2, TEST_Q1, 129, ZERO); - p.add(TEST_FAMILY2, TEST_Q2, 129, ZERO); + p.addColumn(TEST_FAMILY2, TEST_Q1, (long) 129, ZERO); + p.addColumn(TEST_FAMILY2, TEST_Q2, (long) 129, ZERO); p.setACL(permsU1andOwner); t.put(p); } @@ -676,20 +675,20 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { new String[] { user2.getShortName(), AuthUtil.toGroupEntry(GROUP), USER_OWNER.getShortName() }, Action.READ, Action.WRITE); Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 123, ZERO); p.setACL(permsU1andOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 123, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 127, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 127, ZERO); p.setACL(permsU1andOwner); t.put(p); } @@ -768,20 +767,20 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { permsU2andGUandOwner.put(USER_OWNER.getShortName(), new Permission(Permission.Action.READ, Permission.Action.WRITE)); Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 123, ZERO); p.setACL(permsU1andOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 123, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 127, ZERO); p.setACL(permsU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 127, ZERO); p.setACL(permsU1andOwner); t.put(p); } @@ -799,8 +798,8 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { try (Connection connection = ConnectionFactory.createConnection(conf)) { try (Table t = connection.getTable(TEST_TABLE.getTableName())) { Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 125, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 125, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, ZERO); p.setACL(user2.getShortName(), new Permission(Permission.Action.READ, Permission.Action.WRITE)); t.put(p); @@ -864,26 +863,26 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil { AuthUtil.toGroupEntry(GROUP) }, Action.READ, Action.WRITE); Put p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 120, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, 120, ZERO); - p.add(TEST_FAMILY1, TEST_Q3, 120, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 120, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 120, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q3, (long) 120, ZERO); p.setACL(permsU1andU2andGUandOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 123, ZERO); - p.add(TEST_FAMILY1, TEST_Q2, 123, ZERO); - p.add(TEST_FAMILY1, TEST_Q3, 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 123, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q3, (long) 123, ZERO); p.setACL(permsU1andOwner); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q1, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q1, (long) 127, ZERO); p.setACL(permsU1_U2andGU); t.put(p); p = new Put(TEST_ROW1); - p.add(TEST_FAMILY1, TEST_Q2, 127, ZERO); + p.addColumn(TEST_FAMILY1, TEST_Q2, (long) 127, ZERO); p.setACL(user2.getShortName(), new Permission(Permission.Action.READ)); t.put(p); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLs.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLs.java index a283591..7d66507 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLs.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLs.java @@ -161,17 +161,15 @@ public class TestCellACLs extends SecureTestUtil { Table t = connection.getTable(TEST_TABLE.getTableName())) { Put p; // with ro ACL - p = new Put(TEST_ROW).add(TEST_FAMILY, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q1, ZERO); p.setACL(prepareCellPermissions(usersAndGroups, Action.READ)); t.put(p); // with rw ACL - p = new Put(TEST_ROW).add(TEST_FAMILY, TEST_Q2, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q2, ZERO); p.setACL(prepareCellPermissions(usersAndGroups, Action.READ, Action.WRITE)); t.put(p); // no ACL - p = new Put(TEST_ROW) - .add(TEST_FAMILY, TEST_Q3, ZERO) - .add(TEST_FAMILY, TEST_Q4, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q3, ZERO).addColumn(TEST_FAMILY, TEST_Q4, ZERO); t.put(p); } return null; @@ -401,7 +399,7 @@ public class TestCellACLs extends SecureTestUtil { try(Connection connection = ConnectionFactory.createConnection(conf); Table t = connection.getTable(TEST_TABLE.getTableName())) { Put p; - p = new Put(TEST_ROW).add(TEST_FAMILY, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q1, ZERO); t.put(p); } return null; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestScanEarlyTermination.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestScanEarlyTermination.java index 06bc616..74105e1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestScanEarlyTermination.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestScanEarlyTermination.java @@ -161,14 +161,14 @@ public class TestScanEarlyTermination extends SecureTestUtil { Connection connection = ConnectionFactory.createConnection(conf); Table t = connection.getTable(TEST_TABLE.getTableName()); try { - Put put = new Put(TEST_ROW).add(TEST_FAMILY1, TEST_Q1, ZERO); + Put put = new Put(TEST_ROW).addColumn(TEST_FAMILY1, TEST_Q1, ZERO); t.put(put); // Set a READ cell ACL for USER_OTHER on this value in FAMILY2 - put = new Put(TEST_ROW).add(TEST_FAMILY2, TEST_Q1, ZERO); + put = new Put(TEST_ROW).addColumn(TEST_FAMILY2, TEST_Q1, ZERO); put.setACL(USER_OTHER.getShortName(), new Permission(Action.READ)); t.put(put); // Set an empty cell ACL for USER_OTHER on this other value in FAMILY2 - put = new Put(TEST_ROW).add(TEST_FAMILY2, TEST_Q2, ZERO); + put = new Put(TEST_ROW).addColumn(TEST_FAMILY2, TEST_Q2, ZERO); put.setACL(USER_OTHER.getShortName(), new Permission()); t.put(put); } finally { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestTablePermissions.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestTablePermissions.java index 1f23576..c12a47a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestTablePermissions.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestTablePermissions.java @@ -42,7 +42,6 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.HBaseTestingUtility; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.security.User; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -289,10 +288,8 @@ public class TestTablePermissions { AccessControlLists.getTablePermissions(conf, TEST_TABLE); Table table = UTIL.getConnection().getTable(TEST_TABLE); - table.put(new Put(Bytes.toBytes("row1")) - .add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes("v1"))); - table.put(new Put(Bytes.toBytes("row2")) - .add(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes("v2"))); + table.put(new Put(Bytes.toBytes("row1")).addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes("v1"))); + table.put(new Put(Bytes.toBytes("row2")).addColumn(TEST_FAMILY, TEST_QUALIFIER, Bytes.toBytes("v2"))); Admin admin = UTIL.getHBaseAdmin(); admin.split(TEST_TABLE); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java index 880056f..0baead1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java @@ -1010,17 +1010,15 @@ public class TestWithDisabledAuthorization extends SecureTestUtil { Table t = connection.getTable(TEST_TABLE.getTableName())) { Put p; // with ro ACL - p = new Put(TEST_ROW).add(TEST_FAMILY, TEST_Q1, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q1, ZERO); p.setACL(USER_NONE.getShortName(), new Permission(Action.READ)); t.put(p); // with rw ACL - p = new Put(TEST_ROW).add(TEST_FAMILY, TEST_Q2, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q2, ZERO); p.setACL(USER_NONE.getShortName(), new Permission(Action.READ, Action.WRITE)); t.put(p); // no ACL - p = new Put(TEST_ROW) - .add(TEST_FAMILY, TEST_Q3, ZERO) - .add(TEST_FAMILY, TEST_Q4, ZERO); + p = new Put(TEST_ROW).addColumn(TEST_FAMILY, TEST_Q3, ZERO).addColumn(TEST_FAMILY, TEST_Q4, ZERO); t.put(p); } return null; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestDefaultScanLabelGeneratorStack.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestDefaultScanLabelGeneratorStack.java index 39b7f1b..9da2531 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestDefaultScanLabelGeneratorStack.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestDefaultScanLabelGeneratorStack.java @@ -106,15 +106,15 @@ public class TestDefaultScanLabelGeneratorStack { try (Connection connection = ConnectionFactory.createConnection(conf); Table table = TEST_UTIL.createTable(tableName, CF)) { Put put = new Put(ROW_1); - put.add(CF, Q1, HConstants.LATEST_TIMESTAMP, value1); + put.addColumn(CF, Q1, HConstants.LATEST_TIMESTAMP, value1); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q2, HConstants.LATEST_TIMESTAMP, value2); + put.addColumn(CF, Q2, HConstants.LATEST_TIMESTAMP, value2); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q3, HConstants.LATEST_TIMESTAMP, value3); + put.addColumn(CF, Q3, HConstants.LATEST_TIMESTAMP, value3); table.put(put); return null; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestEnforcingScanLabelGenerator.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestEnforcingScanLabelGenerator.java index 3284453..a0703fc 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestEnforcingScanLabelGenerator.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestEnforcingScanLabelGenerator.java @@ -31,7 +31,6 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Table; @@ -104,15 +103,15 @@ public class TestEnforcingScanLabelGenerator { try (Connection connection = ConnectionFactory.createConnection(conf); Table table = TEST_UTIL.createTable(tableName, CF)) { Put put = new Put(ROW_1); - put.add(CF, Q1, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(CF, Q1, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q2, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(CF, Q2, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q3, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(CF, Q3, HConstants.LATEST_TIMESTAMP, value); table.put(put); return null; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java index 047496e..ab2bacc 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabels.java @@ -39,7 +39,6 @@ import org.apache.hadoop.hbase.CellScanner; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; @@ -47,7 +46,6 @@ import org.apache.hadoop.hbase.client.Append; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Increment; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; @@ -577,12 +575,12 @@ public abstract class TestVisibilityLabels { try (Table table = TEST_UTIL.createTable(tableName, fam)) { byte[] row1 = Bytes.toBytes("row1"); Put put = new Put(row1); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(SECRET + " & " + CONFIDENTIAL)); table.checkAndPut(row1, fam, qual, null, put); byte[] row2 = Bytes.toBytes("row2"); put = new Put(row2); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(SECRET)); table.checkAndPut(row2, fam, qual, null, put); @@ -604,7 +602,7 @@ public abstract class TestVisibilityLabels { byte[] row1 = Bytes.toBytes("row1"); byte[] val = Bytes.toBytes(1L); Put put = new Put(row1); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, val); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, val); put.setCellVisibility(new CellVisibility(SECRET + " & " + CONFIDENTIAL)); table.put(put); Get get = new Get(row1); @@ -630,7 +628,7 @@ public abstract class TestVisibilityLabels { byte[] row1 = Bytes.toBytes("row1"); byte[] val = Bytes.toBytes("a"); Put put = new Put(row1); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, val); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, val); put.setCellVisibility(new CellVisibility(SECRET + " & " + CONFIDENTIAL)); table.put(put); Get get = new Get(row1); @@ -710,32 +708,32 @@ public abstract class TestVisibilityLabels { TEST_UTIL.getHBaseAdmin().createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(r1); - put.add(fam, qual, 3l, v1); - put.add(fam, qual2, 3l, v1); - put.add(fam2, qual, 3l, v1); - put.add(fam2, qual2, 3l, v1); + put.addColumn(fam, qual, 3l, v1); + put.addColumn(fam, qual2, 3l, v1); + put.addColumn(fam2, qual, 3l, v1); + put.addColumn(fam2, qual2, 3l, v1); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); put = new Put(r1); - put.add(fam, qual, 4l, v2); - put.add(fam, qual2, 4l, v2); - put.add(fam2, qual, 4l, v2); - put.add(fam2, qual2, 4l, v2); + put.addColumn(fam, qual, 4l, v2); + put.addColumn(fam, qual2, 4l, v2); + put.addColumn(fam2, qual, 4l, v2); + put.addColumn(fam2, qual2, 4l, v2); put.setCellVisibility(new CellVisibility(PRIVATE)); table.put(put); put = new Put(r2); - put.add(fam, qual, 3l, v1); - put.add(fam, qual2, 3l, v1); - put.add(fam2, qual, 3l, v1); - put.add(fam2, qual2, 3l, v1); + put.addColumn(fam, qual, 3l, v1); + put.addColumn(fam, qual2, 3l, v1); + put.addColumn(fam2, qual, 3l, v1); + put.addColumn(fam2, qual2, 3l, v1); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); put = new Put(r2); - put.add(fam, qual, 4l, v2); - put.add(fam, qual2, 4l, v2); - put.add(fam2, qual, 4l, v2); - put.add(fam2, qual2, 4l, v2); + put.addColumn(fam, qual, 4l, v2); + put.addColumn(fam, qual2, 4l, v2); + put.addColumn(fam2, qual, 4l, v2); + put.addColumn(fam2, qual2, 4l, v2); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); @@ -794,11 +792,11 @@ public abstract class TestVisibilityLabels { TEST_UTIL.getHBaseAdmin().createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)){ Put p1 = new Put(row1); - p1.add(fam, qual, value); + p1.addColumn(fam, qual, value); p1.setCellVisibility(new CellVisibility(CONFIDENTIAL)); Put p2 = new Put(row1); - p2.add(fam, qual2, value); + p2.addColumn(fam, qual2, value); p2.setCellVisibility(new CellVisibility(SECRET)); RowMutations rm = new RowMutations(row1); @@ -830,11 +828,11 @@ public abstract class TestVisibilityLabels { TEST_UTIL.getHBaseAdmin().createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put p1 = new Put(row1); - p1.add(fam, qual, value); + p1.addColumn(fam, qual, value); p1.setCellVisibility(new CellVisibility(CONFIDENTIAL)); Put p2 = new Put(row1); - p2.add(fam, qual2, value); + p2.addColumn(fam, qual2, value); p2.setCellVisibility(new CellVisibility(SECRET)); RowMutations rm = new RowMutations(row1); @@ -858,7 +856,7 @@ public abstract class TestVisibilityLabels { List puts = new ArrayList(); for (int i = 0; i < labelExps.length; i++) { Put put = new Put(Bytes.toBytes("row" + (i+1))); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(labelExps[i])); puts.add(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsReplication.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsReplication.java index 7fa240e..00186c1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsReplication.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsReplication.java @@ -381,7 +381,7 @@ public class TestVisibilityLabelsReplication { List puts = new ArrayList(); for (String labelExp : labelExps) { Put put = new Put(Bytes.toBytes("row" + i)); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(labelExp)); put.setAttribute(NON_VISIBILITY, Bytes.toBytes(TEMP)); puts.add(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java index a6192fc..f67296d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithACL.java @@ -318,7 +318,7 @@ public class TestVisibilityLabelsWithACL { List puts = new ArrayList(); for (String labelExp : labelExps) { Put put = new Put(Bytes.toBytes("row" + i)); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(labelExp)); puts.add(put); i++; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.java index 798139c..52c43b2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.java @@ -480,11 +480,11 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, value); + put.addColumn(fam, qual, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, value); + put.addColumn(fam, qual, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -511,7 +511,7 @@ public class TestVisibilityLabelsWithDeletes { Result[] next = scanner.next(3); assertEquals(next.length, 1); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, value1); + put.addColumn(fam, qual, value1); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); actiona = new PrivilegedExceptionAction() { @@ -556,11 +556,11 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, value); + put.addColumn(fam, qual, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, value); + put.addColumn(fam, qual, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -587,7 +587,7 @@ public class TestVisibilityLabelsWithDeletes { Result[] next = scanner.next(3); assertEquals(next.length, 1); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, value1); + put.addColumn(fam, qual, value1); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); actiona = new PrivilegedExceptionAction() { @@ -632,11 +632,11 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 123l, value); + put.addColumn(fam, qual, 123l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 124l, value1); + put.addColumn(fam, qual, 124l, value1); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); PrivilegedExceptionAction actiona = new PrivilegedExceptionAction() { @@ -686,11 +686,11 @@ public class TestVisibilityLabelsWithDeletes { try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put1 = new Put(Bytes.toBytes("row1")); - put1.add(fam, qual, 123l, value); + put1.addColumn(fam, qual, 123l, value); put1.setCellVisibility(new CellVisibility(CONFIDENTIAL)); Put put2 = new Put(Bytes.toBytes("row1")); - put2.add(fam, qual, 123l, value1); + put2.addColumn(fam, qual, 123l, value1); put2.setCellVisibility(new CellVisibility(SECRET)); table.put(createList(put1, put2)); @@ -861,36 +861,36 @@ public class TestVisibilityLabelsWithDeletes { List puts = new ArrayList(); Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 123l, value); + put.addColumn(fam, qual, 123l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 124l, value); + put.addColumn(fam, qual, 124l, value); put.setCellVisibility(new CellVisibility("(" + CONFIDENTIAL + "&" + PRIVATE + ")|(" + TOPSECRET + "&" + SECRET+")")); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 125l, value); + put.addColumn(fam, qual, 125l, value); put.setCellVisibility(new CellVisibility(SECRET + "&" + TOPSECRET)); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 126l, value); + put.addColumn(fam, qual, 126l, value); put.setCellVisibility(new CellVisibility("(" + CONFIDENTIAL + "&" + PRIVATE + ")|(" + TOPSECRET + "&" + SECRET+")")); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 127l, value); + put.addColumn(fam, qual, 127l, value); put.setCellVisibility(new CellVisibility("(" + CONFIDENTIAL + "&" + PRIVATE + ")|(" + TOPSECRET + "&" + SECRET+")")); puts.add(put); TEST_UTIL.getHBaseAdmin().flush(tableName); put = new Put(Bytes.toBytes("row2")); - put.add(fam, qual, 127l, value); + put.addColumn(fam, qual, 127l, value); put.setCellVisibility(new CellVisibility("(" + CONFIDENTIAL + "&" + PRIVATE + ")|(" + TOPSECRET + "&" + SECRET + ")")); puts.add(put); @@ -911,28 +911,28 @@ public class TestVisibilityLabelsWithDeletes { List puts = new ArrayList<>(); Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 123l, value); + put.addColumn(fam, qual, 123l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 124l, value); + put.addColumn(fam, qual, 124l, value); put.setCellVisibility(new CellVisibility("(" + CONFIDENTIAL + "&" + PRIVATE + ")|(" + TOPSECRET + "&" + SECRET+")")); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 125l, value); + put.addColumn(fam, qual, 125l, value); put.setCellVisibility(new CellVisibility(SECRET + "&" + TOPSECRET)); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual1, 126l, value); + put.addColumn(fam, qual1, 126l, value); put.setCellVisibility(new CellVisibility(SECRET + "&" + TOPSECRET)); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual2, 127l, value); + put.addColumn(fam, qual2, 127l, value); put.setCellVisibility(new CellVisibility("(" + CONFIDENTIAL + "&" + PRIVATE + ")|(" + TOPSECRET + "&" + SECRET+")")); puts.add(put); @@ -952,23 +952,23 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); List puts = new ArrayList<>(); Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 123l, value); + put.addColumn(fam, qual, 123l, value); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 124l, value); + put.addColumn(fam, qual, 124l, value); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 125l, value); + put.addColumn(fam, qual, 125l, value); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 126l, value); + put.addColumn(fam, qual, 126l, value); puts.add(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 127l, value); + put.addColumn(fam, qual, 127l, value); puts.add(put); Table table = TEST_UTIL.getConnection().getTable(tableName); @@ -977,7 +977,7 @@ public class TestVisibilityLabelsWithDeletes { TEST_UTIL.getHBaseAdmin().flush(tableName); put = new Put(Bytes.toBytes("row2")); - put.add(fam, qual, 127l, value); + put.addColumn(fam, qual, 127l, value); table.put(put); return table; @@ -1116,7 +1116,7 @@ public class TestVisibilityLabelsWithDeletes { try (Table table = doPuts(tableName)) { TEST_UTIL.getHBaseAdmin().flush(tableName); Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 128l, value); + put.addColumn(fam, qual, 128l, value); put.setCellVisibility(new CellVisibility(TOPSECRET)); table.put(put); PrivilegedExceptionAction actiona = new PrivilegedExceptionAction() { @@ -1176,7 +1176,7 @@ public class TestVisibilityLabelsWithDeletes { current.getRowLength(), row2, 0, row2.length)); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 129l, value); + put.addColumn(fam, qual, 129l, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); @@ -1220,7 +1220,7 @@ public class TestVisibilityLabelsWithDeletes { SUPERUSER.runAs(actiona); TEST_UTIL.getHBaseAdmin().flush(tableName); Put put = new Put(Bytes.toBytes("row3")); - put.add(fam, qual, 127l, value); + put.addColumn(fam, qual, 127l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL + "&" + PRIVATE)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -1380,11 +1380,11 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual1, 125l, value); + put.addColumn(fam, qual1, 125l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual1, 126l, value); + put.addColumn(fam, qual1, 126l, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -1428,11 +1428,11 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual1, 125l, value); + put.addColumn(fam, qual1, 125l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual1, 126l, value); + put.addColumn(fam, qual1, 126l, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -1647,7 +1647,7 @@ public class TestVisibilityLabelsWithDeletes { TEST_UTIL.getHBaseAdmin().flush(tableName); Put put = new Put(Bytes.toBytes("row3")); - put.add(fam, qual, 127l, value); + put.addColumn(fam, qual, 127l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL + "&" + PRIVATE)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -2782,11 +2782,11 @@ public class TestVisibilityLabelsWithDeletes { hBaseAdmin.createTable(desc); try (Table table = TEST_UTIL.getConnection().getTable(tableName)) { Put put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 123l, value); + put.addColumn(fam, qual, 123l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(Bytes.toBytes("row1")); - put.add(fam, qual, 124l, value); + put.addColumn(fam, qual, 124l, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL + "|" + PRIVATE)); table.put(put); TEST_UTIL.getHBaseAdmin().flush(tableName); @@ -2941,7 +2941,7 @@ public class TestVisibilityLabelsWithDeletes { List puts = new ArrayList(); for (String labelExp : labelExps) { Put put = new Put(Bytes.toBytes("row" + i)); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(labelExp)); puts.add(put); table.put(put); @@ -2959,7 +2959,7 @@ public class TestVisibilityLabelsWithDeletes { List puts = new ArrayList(); for (String labelExp : labelExps) { Put put = new Put(Bytes.toBytes("row" + i)); - put.add(fam, qual, timestamp[i - 1], value); + put.addColumn(fam, qual, timestamp[i - 1], value); put.setCellVisibility(new CellVisibility(labelExp)); puts.add(put); table.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.java index c48b4c4..b6a1c6d 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.java @@ -86,11 +86,11 @@ public class TestVisibilityLabelsWithSLGStack { TableName tableName = TableName.valueOf(TEST_NAME.getMethodName()); try (Table table = TEST_UTIL.createTable(tableName, CF)) { Put put = new Put(ROW_1); - put.add(CF, Q1, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(CF, Q1, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q2, HConstants.LATEST_TIMESTAMP, value); + put.addColumn(CF, Q2, HConstants.LATEST_TIMESTAMP, value); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java index 6c1a47b..1410c78 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityLablesWithGroups.java @@ -120,15 +120,15 @@ public class TestVisibilityLablesWithGroups { try (Connection connection = ConnectionFactory.createConnection(conf); Table table = connection.getTable(tableName)) { Put put = new Put(ROW_1); - put.add(CF, Q1, HConstants.LATEST_TIMESTAMP, value1); + put.addColumn(CF, Q1, HConstants.LATEST_TIMESTAMP, value1); put.setCellVisibility(new CellVisibility(SECRET)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q2, HConstants.LATEST_TIMESTAMP, value2); + put.addColumn(CF, Q2, HConstants.LATEST_TIMESTAMP, value2); put.setCellVisibility(new CellVisibility(CONFIDENTIAL)); table.put(put); put = new Put(ROW_1); - put.add(CF, Q3, HConstants.LATEST_TIMESTAMP, value3); + put.addColumn(CF, Q3, HConstants.LATEST_TIMESTAMP, value3); table.put(put); } return null; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityWithCheckAuths.java hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityWithCheckAuths.java index 0b0d9d0..96203b5 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityWithCheckAuths.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/security/visibility/TestVisibilityWithCheckAuths.java @@ -137,7 +137,7 @@ public class TestVisibilityWithCheckAuths { Table table = connection.getTable(tableName)) { Put p = new Put(row1); p.setCellVisibility(new CellVisibility(PUBLIC + "&" + TOPSECRET)); - p.add(fam, qual, 125l, value); + p.addColumn(fam, qual, 125l, value); table.put(p); Assert.fail("Testcase should fail with AccesDeniedException"); } catch (Throwable t) { @@ -177,7 +177,7 @@ public class TestVisibilityWithCheckAuths { try (Connection connection = ConnectionFactory.createConnection(conf); Table table = connection.getTable(tableName)) { Put put = new Put(row1); - put.add(fam, qual, HConstants.LATEST_TIMESTAMP, val); + put.addColumn(fam, qual, HConstants.LATEST_TIMESTAMP, val); put.setCellVisibility(new CellVisibility(TOPSECRET)); table.put(put); } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java index 6c58d62..39c8f21 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java @@ -44,12 +44,9 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotEnabledException; -import org.apache.hadoop.hbase.Waiter; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.BufferedMutator; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher; @@ -713,7 +710,7 @@ public class SnapshotTestingUtils { Put put = new Put(key); put.setDurability(Durability.SKIP_WAL); for (byte[] family: families) { - put.add(family, q, value); + put.addColumn(family, q, value); } return put; } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/trace/TestHTraceHooks.java hbase-server/src/test/java/org/apache/hadoop/hbase/trace/TestHTraceHooks.java index 65dff1b..205f1d8 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/trace/TestHTraceHooks.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/trace/TestHTraceHooks.java @@ -42,8 +42,6 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; -import com.google.common.collect.Multimap; - @Category({MiscTests.class, MediumTests.class}) public class TestHTraceHooks { @@ -118,7 +116,7 @@ public class TestHTraceHooks { assertTrue(spans.size() > 5); Put put = new Put("row".getBytes()); - put.add(FAMILY_BYTES, "col".getBytes(), "value".getBytes()); + put.addColumn(FAMILY_BYTES, "col".getBytes(), "value".getBytes()); TraceScope putSpan = Trace.startSpan("doing put", Sampler.ALWAYS); try { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/BaseTestHBaseFsck.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/BaseTestHBaseFsck.java index cc29732..8662739 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/BaseTestHBaseFsck.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/BaseTestHBaseFsck.java @@ -30,28 +30,16 @@ import static org.junit.Assert.fail; import java.io.IOException; import java.util.ArrayList; -import java.util.Arrays; import java.util.Collection; import java.util.HashMap; -import java.util.HashSet; -import java.util.LinkedList; import java.util.List; import java.util.Map; -import java.util.NavigableMap; -import java.util.Set; import java.util.UUID; -import java.util.concurrent.Callable; import java.util.concurrent.CountDownLatch; import java.util.concurrent.ExecutorService; -import java.util.concurrent.Executors; -import java.util.concurrent.Future; import java.util.concurrent.ScheduledThreadPoolExecutor; -import java.util.concurrent.SynchronousQueue; -import java.util.concurrent.ThreadPoolExecutor; -import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicBoolean; -import org.apache.commons.io.IOUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -74,28 +62,19 @@ import org.apache.hadoop.hbase.client.ClusterConnection; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Delete; -import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; -import org.apache.hadoop.hbase.client.RegionReplicaUtil; -import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.coprocessor.BaseMasterObserver; -import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; import org.apache.hadoop.hbase.coprocessor.MasterCoprocessorEnvironment; import org.apache.hadoop.hbase.coprocessor.ObserverContext; -import org.apache.hadoop.hbase.io.hfile.TestHFile; import org.apache.hadoop.hbase.master.AssignmentManager; -import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.RegionState; import org.apache.hadoop.hbase.master.RegionStates; -import org.apache.hadoop.hbase.master.TableLockManager; -import org.apache.hadoop.hbase.master.TableLockManager.TableLock; import org.apache.hadoop.hbase.mob.MobFileName; import org.apache.hadoop.hbase.mob.MobUtils; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; @@ -105,29 +84,16 @@ import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.SplitTransactionFactory; import org.apache.hadoop.hbase.regionserver.SplitTransactionImpl; -import org.apache.hadoop.hbase.regionserver.TestEndToEndSplitTransaction; -import org.apache.hadoop.hbase.testclassification.LargeTests; -import org.apache.hadoop.hbase.testclassification.MiscTests; import org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter; import org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE; import org.apache.hadoop.hbase.util.HBaseFsck.HbckInfo; -import org.apache.hadoop.hbase.util.HBaseFsck.PrintingErrorReporter; import org.apache.hadoop.hbase.util.HBaseFsck.TableInfo; import org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker; -import org.apache.hadoop.hbase.util.hbck.HbckTestingUtil; -import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; import org.apache.zookeeper.KeeperException; -import org.junit.AfterClass; -import org.junit.Assert; -import org.junit.Before; -import org.junit.BeforeClass; import org.junit.Ignore; import org.junit.Test; -import org.junit.experimental.categories.Category; import org.junit.rules.TestName; -import com.google.common.collect.Multimap; - /** * This is the base class for HBaseFsck's ability to detect reasons for inconsistent tables. * @@ -342,7 +308,7 @@ public class BaseTestHBaseFsck { List puts = new ArrayList(); for (byte[] row : ROWKEYS) { Put p = new Put(row); - p.add(FAM, Bytes.toBytes("val"), row); + p.addColumn(FAM, Bytes.toBytes("val"), row); puts.add(p); } tbl.put(puts); @@ -497,7 +463,7 @@ public class BaseTestHBaseFsck { tbl = connection.getTable(desc.getTableName()); for (int i = 0; i < 5; i++) { Put p1 = new Put(("r" + i).getBytes()); - p1.add(Bytes.toBytes("f"), "q1".getBytes(), "v".getBytes()); + p1.addColumn(Bytes.toBytes("f"), "q1".getBytes(), "v".getBytes()); tbl.put(p1); } admin.flush(desc.getTableName()); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedUpdater.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedUpdater.java index f2b82a4..e28acc6 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedUpdater.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedUpdater.java @@ -208,7 +208,7 @@ public class MultiThreadedUpdater extends MultiThreadedWriterBase { switch (mt) { case PUT: Put put = new Put(rowKey); - put.add(cf, column, hashCodeBytes); + put.addColumn(cf, column, hashCodeBytes); mutate(table, put, rowKeyBase, rowKey, cf, column, checkedValue); buf.append(MutationType.PUT.getNumber()); break; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedWriter.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedWriter.java index 4d1c286..83e207a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedWriter.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedWriter.java @@ -105,7 +105,7 @@ public class MultiThreadedWriter extends MultiThreadedWriterBase { byte[][] columns = dataGenerator.generateColumnsForCf(rowKey, cf); for (byte[] column : columns) { byte[] value = dataGenerator.generateValue(rowKey, cf, column); - put.add(cf, column, value); + put.addColumn(cf, column, value); ++columnCount; if (!isMultiPut) { insert(table, put, rowKeyBase); @@ -114,8 +114,8 @@ public class MultiThreadedWriter extends MultiThreadedWriterBase { } } long rowKeyHash = Arrays.hashCode(rowKey); - put.add(cf, MUTATE_INFO, HConstants.EMPTY_BYTE_ARRAY); - put.add(cf, INCREMENT, Bytes.toBytes(rowKeyHash)); + put.addColumn(cf, MUTATE_INFO, HConstants.EMPTY_BYTE_ARRAY); + put.addColumn(cf, INCREMENT, Bytes.toBytes(rowKeyHash)); if (!isMultiPut) { insert(table, put, rowKeyBase); numCols.addAndGet(1); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java index 451ccee..445914f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java @@ -110,17 +110,17 @@ public class TestCoprocessorScanPolicy { // set the version override to 2 Put p = new Put(R); p.setAttribute("versions", new byte[]{}); - p.add(F, tableName.getName(), Bytes.toBytes(2)); + p.addColumn(F, tableName.getName(), Bytes.toBytes(2)); t.put(p); long now = EnvironmentEdgeManager.currentTime(); // insert 2 versions p = new Put(R); - p.add(F, Q, now, Q); + p.addColumn(F, Q, now, Q); t.put(p); p = new Put(R); - p.add(F, Q, now+1, Q); + p.addColumn(F, Q, now + 1, Q); t.put(p); Get g = new Get(R); g.setMaxVersions(10); @@ -138,7 +138,7 @@ public class TestCoprocessorScanPolicy { // insert a 3rd version p = new Put(R); - p.add(F, Q, now+2, Q); + p.addColumn(F, Q, now + 2, Q); t.put(p); g = new Get(R); g.setMaxVersions(10); @@ -172,14 +172,14 @@ public class TestCoprocessorScanPolicy { // Set the TTL override to 3s Put p = new Put(R); p.setAttribute("ttl", new byte[]{}); - p.add(F, tableName.getName(), Bytes.toBytes(3000L)); + p.addColumn(F, tableName.getName(), Bytes.toBytes(3000L)); t.put(p); p = new Put(R); - p.add(F, Q, ts, Q); + p.addColumn(F, Q, ts, Q); t.put(p); p = new Put(R); - p.add(F, Q, ts+1, Q); + p.addColumn(F, Q, ts + 1, Q); t.put(p); // these two should be expired but for the override diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java index 5c08f1f..0ed8711 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java @@ -109,8 +109,8 @@ public class TestHBaseFsckEncryption { for (int i = 0; i < values.length; i++) { for (int j = 0; j < values.length; j++) { Put put = new Put(new byte[] { values[i], values[j] }); - put.add(Bytes.toBytes("cf"), new byte[] {}, new byte[] { values[i], - values[j] }); + put.addColumn(Bytes.toBytes("cf"), new byte[]{}, new byte[]{values[i], + values[j]}); table.put(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java index a44ccd1..b833a6a 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java @@ -1359,10 +1359,8 @@ public class TestHBaseFsckOneRS extends BaseTestHBaseFsck { meta.delete(deletes); // Mess it up by creating a fake hbase:meta entry with no associated RegionInfo - meta.put(new Put(Bytes.toBytes(table + ",,1361911384013.810e28f59a57da91c66")).add( - HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes.toBytes("node1:60020"))); - meta.put(new Put(Bytes.toBytes(table + ",,1361911384013.810e28f59a57da91c66")).add( - HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(1362150791183L))); + meta.put(new Put(Bytes.toBytes(table + ",,1361911384013.810e28f59a57da91c66")).addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes.toBytes("node1:60020"))); + meta.put(new Put(Bytes.toBytes(table + ",,1361911384013.810e28f59a57da91c66")).addColumn(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(1362150791183L))); meta.close(); HBaseFsck hbck = doFsck(conf, false); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckTwoRS.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckTwoRS.java index 4eb1dd8..8679060 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckTwoRS.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckTwoRS.java @@ -363,8 +363,7 @@ public class TestHBaseFsckTwoRS extends BaseTestHBaseFsck { assertNotNull(serverName); try (Table meta = connection.getTable(TableName.META_TABLE_NAME, tableExecutorService)) { Put put = new Put(regionName); - put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, - Bytes.toBytes(serverName.getHostAndPort())); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes.toBytes(serverName.getHostAndPort())); meta.put(put); } @@ -432,10 +431,8 @@ public class TestHBaseFsckTwoRS extends BaseTestHBaseFsck { startCode != sn.getStartcode()) { Put put = new Put(res.getRow()); put.setDurability(Durability.SKIP_WAL); - put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, - Bytes.toBytes(sn.getHostAndPort())); - put.add(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, - Bytes.toBytes(sn.getStartcode())); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER, Bytes.toBytes(sn.getHostAndPort())); + put.addColumn(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn.getStartcode())); meta.put(put); hri = MetaTableAccessor.getHRegionInfo(res); break; diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java index e31cb4e..661af14 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTable.java @@ -34,7 +34,6 @@ import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.regionserver.HRegion; -import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.MiscTests; import org.junit.Test; @@ -147,7 +146,7 @@ public class TestMergeTable { for(int i = firstRow; i < firstRow + nrows; i++) { Put put = new Put(Bytes.toBytes("row_" + String.format("%1$05d", i))); put.setDurability(Durability.SKIP_WAL); - put.add(COLUMN_NAME, null, VALUE); + put.addColumn(COLUMN_NAME, null, VALUE); region.put(put); if (i % 10000 == 0) { LOG.info("Flushing write #" + i); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java index a1b2aa1..cb51fb2 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java @@ -164,7 +164,7 @@ public class TestMergeTool extends HBaseTestCase { for (int j = 0; j < rows[i].length; j++) { byte [] row = rows[i][j]; Put put = new Put(row); - put.add(FAMILY, QUALIFIER, row); + put.addColumn(FAMILY, QUALIFIER, row); regions[i].put(put); } HRegion.addRegionToMETA(meta, regions[i]); diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRebuildTestCore.java hbase-server/src/test/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRebuildTestCore.java index 578ec85..ecda77f 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRebuildTestCore.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRebuildTestCore.java @@ -22,8 +22,6 @@ import static org.junit.Assert.assertEquals; import java.io.IOException; import java.util.ArrayList; import java.util.List; -import java.util.Map; -import java.util.Map.Entry; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -45,7 +43,6 @@ import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Result; @@ -153,8 +150,8 @@ public class OfflineMetaRebuildTestCore { for (int i = 0; i < values.length; i++) { for (int j = 0; j < values.length; j++) { Put put = new Put(new byte[] { values[i], values[j] }); - put.add(Bytes.toBytes("fam"), new byte[] {}, new byte[] { values[i], - values[j] }); + put.addColumn(Bytes.toBytes("fam"), new byte[]{}, new byte[]{values[i], + values[j]}); puts.add(put); } } diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java index 8833eda..bde3e49 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFiltering.java @@ -89,7 +89,7 @@ public class TestWALFiltering { final byte[] value = Bytes.toBytes("value_for_row_" + iRow + "_cf_" + Bytes.toStringBinary(cf) + "_col_" + iCol + "_ts_" + ts + "_random_" + rand.nextLong()); - put.add(cf, qual, ts, value); + put.addColumn(cf, qual, ts, value); } else if (rand.nextDouble() < 0.8) { del.addColumn(cf, qual, ts); } else { diff --git hbase-server/src/test/java/org/apache/hadoop/hbase/wal/WALPerformanceEvaluation.java hbase-server/src/test/java/org/apache/hadoop/hbase/wal/WALPerformanceEvaluation.java index 3af853b..6866df1 100644 --- hbase-server/src/test/java/org/apache/hadoop/hbase/wal/WALPerformanceEvaluation.java +++ hbase-server/src/test/java/org/apache/hadoop/hbase/wal/WALPerformanceEvaluation.java @@ -545,7 +545,7 @@ public final class WALPerformanceEvaluation extends Configured implements Tool { for (int cf = 0; cf < numFamilies; ++cf) { for (int q = 0; q < numQualifiers; ++q) { rand.nextBytes(value); - put.add(Bytes.toBytes(FAMILY_PREFIX + cf), Bytes.toBytes(QUALIFIER_PREFIX + q), value); + put.addColumn(Bytes.toBytes(FAMILY_PREFIX + cf), Bytes.toBytes(QUALIFIER_PREFIX + q), value); } } return put;