From 048bb6e1b3866724b1b980514cd0a864ef855f28 Mon Sep 17 00:00:00 2001 From: Jan Hentschel Date: Sun, 17 Dec 2017 17:27:58 +0100 Subject: [PATCH] HBASE-19540 Reduced number of unnecessary semicolons --- .../org/apache/hadoop/hbase/io/MetricsIOSourceImpl.java | 5 ++--- .../hadoop/hbase/IntegrationTestBackupRestore.java | 3 +-- .../hadoop/hbase/IntegrationTestIngestWithMOB.java | 2 +- .../hbase/StripeCompactionsPerformanceEvaluation.java | 2 +- .../actions/RollingBatchRestartRsExceptMetaAction.java | 3 +-- .../IntegrationTestWithCellVisibilityLoadAndVerify.java | 2 +- .../org/apache/hadoop/hbase/mapreduce/SyncTable.java | 2 +- .../org/apache/hadoop/hbase/PerformanceEvaluation.java | 2 +- .../apache/hadoop/hbase/mapreduce/TestCopyTable.java | 2 +- .../apache/hadoop/hbase/mapreduce/TestImportExport.java | 17 +++++++---------- .../hbase/procedure2/RemoteProcedureDispatcher.java | 2 +- .../hbase/procedure2/store/wal/WALProcedureStore.java | 2 +- .../hadoop/hbase/procedure2/TestProcedureToString.java | 2 +- .../hbase/procedure2/TestStateMachineProcedure.java | 3 ++- .../apache/hadoop/hbase/rest/model/ScannerModel.java | 2 +- .../hadoop/hbase/rsgroup/RSGroupBasedLoadBalancer.java | 2 +- .../apache/hadoop/hbase/thrift/TestThriftServer.java | 2 +- 17 files changed, 25 insertions(+), 30 deletions(-) diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.java index 3edbc55e47..b41a46f7b4 100644 --- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.java +++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.java @@ -56,12 +56,12 @@ public class MetricsIOSourceImpl extends BaseSourceImpl implements MetricsIOSour @Override public void updateFsReadTime(long t) { fsReadTimeHisto.add(t); - }; + } @Override public void updateFsPReadTime(long t) { fsPReadTimeHisto.add(t); - }; + } @Override public void updateFsWriteTime(long t) { @@ -80,5 +80,4 @@ public class MetricsIOSourceImpl extends BaseSourceImpl implements MetricsIOSour metricsRegistry.snapshot(mrb, all); } - } diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java index f041f72f29..de307db0f8 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java @@ -164,8 +164,7 @@ public class IntegrationTestBackupRestore extends IntegrationTestBase { try (Connection conn = util.getConnection(); Admin admin = conn.getAdmin(); - BackupAdmin client = new BackupAdminImpl(conn);) { - + BackupAdmin client = new BackupAdminImpl(conn)) { // #0- insert some data to table TABLE_NAME1, TABLE_NAME2 loadData(TABLE_NAME1, rowsInBatch); loadData(TABLE_NAME2, rowsInBatch); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java index 010e4b90ea..0e0e73ef50 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java @@ -110,7 +110,7 @@ public class IntegrationTestIngestWithMOB extends IntegrationTestIngest { @Test public void testIngest() throws Exception { runIngestTest(JUNIT_RUN_TIME, 100, 10, 1024, 10, 20); - }; + } @Override protected void initTable() throws IOException { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java index 64f7da1355..9e95b01d6c 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java @@ -346,5 +346,5 @@ public class StripeCompactionsPerformanceEvaluation extends AbstractHBaseTool { public boolean verify(byte[] rowKey, byte[] cf, Set columnSet) { return true; } - }; + } } diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java index f03b8ec40e..d0d0fe561d 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java @@ -38,6 +38,5 @@ public class RollingBatchRestartRsExceptMetaAction extends RollingBatchRestartRs List servers = super.selectServers(); servers.remove(metaServer); return servers; - }; - + } } diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java index 3cafe9d433..3f97fbb8cc 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestWithCellVisibilityLoadAndVerify.java @@ -109,7 +109,7 @@ public class IntegrationTestWithCellVisibilityLoadAndVerify extends IntegrationT private static User USER1, USER2; private enum Counters { - ROWS_VIS_EXP_1, ROWS_VIS_EXP_2, ROWS_VIS_EXP_3, ROWS_VIS_EXP_4; + ROWS_VIS_EXP_1, ROWS_VIS_EXP_2, ROWS_VIS_EXP_3, ROWS_VIS_EXP_4 } @Override diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java index edef842f5a..d41f6d9fd7 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java @@ -172,7 +172,7 @@ public class SyncTable extends Configured implements Tool { public static enum Counter {BATCHES, HASHES_MATCHED, HASHES_NOT_MATCHED, SOURCEMISSINGROWS, SOURCEMISSINGCELLS, TARGETMISSINGROWS, TARGETMISSINGCELLS, ROWSWITHDIFFS, DIFFERENTCELLVALUES, - MATCHINGROWS, MATCHINGCELLS, EMPTY_BATCHES, RANGESMATCHED, RANGESNOTMATCHED}; + MATCHINGROWS, MATCHINGCELLS, EMPTY_BATCHES, RANGESMATCHED, RANGESNOTMATCHED} @Override protected void setup(Context context) throws IOException { diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java index 4f8b82f4fa..ef2d6d0c00 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java @@ -1194,7 +1194,7 @@ public class PerformanceEvaluation extends Configured implements Tool { for (int i = startRow; i < lastRow; i++) { if (i % everyN != 0) continue; long startTime = System.nanoTime(); - try (TraceScope scope = TraceUtil.createTrace("test row");){ + try (TraceScope scope = TraceUtil.createTrace("test row")){ testRow(i); } if ( (i - startRow) > opts.measureAfter) { diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java index 0bec03b91d..abb17d6f59 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java @@ -80,7 +80,7 @@ public class TestCopyTable { final byte[] COLUMN1 = Bytes.toBytes("c1"); try (Table t1 = TEST_UTIL.createTable(tableName1, FAMILY); - Table t2 = TEST_UTIL.createTable(tableName2, FAMILY);) { + Table t2 = TEST_UTIL.createTable(tableName2, FAMILY)) { // put rows into the first table for (int i = 0; i < 10; i++) { Put p = new Put(Bytes.toBytes("row" + i)); diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java index fcd01b6f48..e756fc4539 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java @@ -185,7 +185,7 @@ public class TestImportExport { */ @Test public void testSimpleCase() throws Throwable { - try (Table t = UTIL.createTable(TableName.valueOf(name.getMethodName()), FAMILYA, 3);) { + try (Table t = UTIL.createTable(TableName.valueOf(name.getMethodName()), FAMILYA, 3)) { Put p = new Put(ROW1); p.addColumn(FAMILYA, QUAL, now, QUAL); p.addColumn(FAMILYA, QUAL, now + 1, QUAL); @@ -214,7 +214,7 @@ public class TestImportExport { assertTrue(runExport(args)); final String IMPORT_TABLE = name.getMethodName() + "import"; - try (Table t = UTIL.createTable(TableName.valueOf(IMPORT_TABLE), FAMILYB, 3);) { + try (Table t = UTIL.createTable(TableName.valueOf(IMPORT_TABLE), FAMILYB, 3)) { args = new String[] { "-D" + Import.CF_RENAME_PROP + "="+FAMILYA_STRING+":"+FAMILYB_STRING, IMPORT_TABLE, @@ -267,7 +267,7 @@ public class TestImportExport { FileSystem fs = FileSystem.get(UTIL.getConfiguration()); fs.copyFromLocalFile(importPath, new Path(FQ_OUTPUT_DIR + Path.SEPARATOR + name)); String IMPORT_TABLE = name; - try (Table t = UTIL.createTable(TableName.valueOf(IMPORT_TABLE), Bytes.toBytes("f1"), 3);) { + try (Table t = UTIL.createTable(TableName.valueOf(IMPORT_TABLE), Bytes.toBytes("f1"), 3)) { String[] args = new String[] { "-Dhbase.import.version=0.94" , IMPORT_TABLE, FQ_OUTPUT_DIR @@ -297,8 +297,7 @@ public class TestImportExport { .build()) .build(); UTIL.getAdmin().createTable(desc); - try (Table t = UTIL.getConnection().getTable(desc.getTableName());) { - + try (Table t = UTIL.getConnection().getTable(desc.getTableName())) { Put p = new Put(ROW1); p.addColumn(FAMILYA, QUAL, now, QUAL); p.addColumn(FAMILYA, QUAL, now + 1, QUAL); @@ -329,8 +328,7 @@ public class TestImportExport { .build()) .build(); UTIL.getAdmin().createTable(desc); - try (Table t = UTIL.getConnection().getTable(desc.getTableName());) { - + try (Table t = UTIL.getConnection().getTable(desc.getTableName())) { Put p = new Put(ROW1); p.addColumn(FAMILYA, QUAL, now, QUAL); p.addColumn(FAMILYA, QUAL, now + 1, QUAL); @@ -363,7 +361,7 @@ public class TestImportExport { .build()) .build(); UTIL.getAdmin().createTable(desc); - try (Table t = UTIL.getConnection().getTable(desc.getTableName());) { + try (Table t = UTIL.getConnection().getTable(desc.getTableName())) { args = new String[] { IMPORT_TABLE, FQ_OUTPUT_DIR @@ -717,8 +715,7 @@ public class TestImportExport { public void testDurability() throws Throwable { // Create an export table. String exportTableName = name.getMethodName() + "export"; - try (Table exportTable = UTIL.createTable(TableName.valueOf(exportTableName), FAMILYA, 3);) { - + try (Table exportTable = UTIL.createTable(TableName.valueOf(exportTableName), FAMILYA, 3)) { // Insert some data Put put = new Put(ROW1); put.addColumn(FAMILYA, QUAL, now, QUAL); diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java index 2b66e7cf6d..4cee524a12 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java @@ -371,5 +371,5 @@ public abstract class RemoteProcedureDispatcher task, final long delay, final TimeUnit unit) { super(task, EnvironmentEdgeManager.currentTime() + unit.toMillis(delay)); } - }; + } } diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java index 84cda6526f..a3c7cbf423 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java @@ -649,7 +649,7 @@ public class WALProcedureStore extends ProcedureStoreBase { slotsCache.offer(slot); } - private enum PushType { INSERT, UPDATE, DELETE }; + private enum PushType { INSERT, UPDATE, DELETE } private long pushData(final PushType type, final ByteSlot slot, final long procId, final long[] subProcIds) { diff --git a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestProcedureToString.java b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestProcedureToString.java index 9bbef91279..921a8e15b4 100644 --- a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestProcedureToString.java +++ b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestProcedureToString.java @@ -32,7 +32,7 @@ public class TestProcedureToString { /** * A do-nothing environment for BasicProcedure. */ - static class BasicProcedureEnv {}; + static class BasicProcedureEnv {} /** * A do-nothing basic procedure just for testing toString. diff --git a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestStateMachineProcedure.java b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestStateMachineProcedure.java index cbe50f2c2d..a61370c544 100644 --- a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestStateMachineProcedure.java +++ b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/TestStateMachineProcedure.java @@ -149,7 +149,8 @@ public class TestStateMachineProcedure { assertEquals(TEST_FAILURE_EXCEPTION, cause); } - public enum TestSMProcedureState { STEP_1, STEP_2 }; + public enum TestSMProcedureState { STEP_1, STEP_2 } + public static class TestSMProcedure extends StateMachineProcedure { protected Flow executeFromState(TestProcEnv env, TestSMProcedureState state) { diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java index a678fdede4..7748272aa8 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java @@ -107,7 +107,7 @@ public class ScannerModel implements ProtobufMessageHandler, Serializable { private static final long serialVersionUID = 1L; private byte[] startRow = HConstants.EMPTY_START_ROW; - private byte[] endRow = HConstants.EMPTY_END_ROW;; + private byte[] endRow = HConstants.EMPTY_END_ROW; private List columns = new ArrayList<>(); private int batch = Integer.MAX_VALUE; private long startTime = 0; diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupBasedLoadBalancer.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupBasedLoadBalancer.java index d838edb94c..0c24ce51e9 100644 --- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupBasedLoadBalancer.java +++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupBasedLoadBalancer.java @@ -209,7 +209,7 @@ public class RSGroupBasedLoadBalancer implements RSGroupableBalancer { } for (RegionInfo region : misplacedRegions) { - String groupName = rsGroupInfoManager.getRSGroupOfTable(region.getTable());; + String groupName = rsGroupInfoManager.getRSGroupOfTable(region.getTable()); RSGroupInfo info = rsGroupInfoManager.getRSGroup(groupName); List candidateList = filterOfflineServers(info, servers); ServerName server = this.internalBalancer.randomAssignment(region, diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java index d0052e50be..d19e336269 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java @@ -210,7 +210,7 @@ public class TestThriftServer { int currentCountDeleteTable = getCurrentCount("deleteTable_num_ops", 2, metrics); int currentCountDisableTable = getCurrentCount("disableTable_num_ops", 2, metrics); createTestTables(handler); - dropTestTables(handler);; + dropTestTables(handler); metricsHelper.assertCounter("createTable_num_ops", currentCountCreateTable + 2, metrics.getSource()); metricsHelper.assertCounter("deleteTable_num_ops", currentCountDeleteTable + 2, -- 2.14.3 (Apple Git-98)