From 935fb3f518b63d2e07d9ab7ecda525243d330a11 Mon Sep 17 00:00:00 2001 From: Vladimir Rodionov Date: Wed, 17 Jan 2018 16:34:18 -0800 Subject: [PATCH] HBASE-17852: Add Fault tolerance to HBASE-14417 (Support bulk loaded files in incremental backup) --- .../hadoop/hbase/backup/BackupHFileCleaner.java | 6 +- .../apache/hadoop/hbase/backup/BackupObserver.java | 6 +- .../apache/hadoop/hbase/backup/RestoreDriver.java | 4 +- .../hadoop/hbase/backup/impl/BackupAdminImpl.java | 46 +++---- .../hadoop/hbase/backup/impl/BackupCommands.java | 28 ++-- .../hadoop/hbase/backup/impl/BackupManager.java | 8 +- ...BackupSystemTable.java => BackupMetaTable.java} | 148 ++++++++++----------- .../backup/impl/IncrementalBackupManager.java | 2 +- .../hbase/backup/impl/TableBackupClient.java | 8 +- .../backup/mapreduce/MapReduceBackupMergeJob.java | 8 +- .../hbase/backup/master/BackupLogCleaner.java | 4 +- .../regionserver/LogRollBackupSubprocedure.java | 4 +- .../apache/hadoop/hbase/backup/TestBackupBase.java | 4 +- .../hadoop/hbase/backup/TestBackupDelete.java | 4 +- .../hbase/backup/TestBackupDeleteWithFailures.java | 4 +- .../hadoop/hbase/backup/TestBackupDescribe.java | 4 +- .../hbase/backup/TestBackupHFileCleaner.java | 4 +- ...upSystemTable.java => TestBackupMetaTable.java} | 34 ++--- .../hadoop/hbase/backup/TestBackupRepair.java | 4 +- .../apache/hadoop/hbase/backup/TestFullBackup.java | 4 +- .../hadoop/hbase/backup/TestFullBackupSet.java | 4 +- .../hbase/backup/TestFullBackupSetRestoreSet.java | 6 +- .../hbase/backup/TestFullBackupWithFailures.java | 4 +- .../TestIncrementalBackupMergeWithFailures.java | 6 +- .../backup/TestIncrementalBackupWithBulkLoad.java | 4 +- .../backup/TestIncrementalBackupWithFailures.java | 4 +- .../hbase/backup/TestRepairAfterFailedDelete.java | 10 +- .../hbase/backup/TestSystemTableSnapshot.java | 4 +- .../hbase/backup/master/TestBackupLogCleaner.java | 4 +- .../hadoop/hbase/IntegrationTestBackupRestore.java | 4 +- 30 files changed, 192 insertions(+), 192 deletions(-) rename hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/{BackupSystemTable.java => BackupMetaTable.java} (92%) rename hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/{TestBackupSystemTable.java => TestBackupMetaTable.java} (94%) diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupHFileCleaner.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupHFileCleaner.java index da4f56e9d3..ef168d00d0 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupHFileCleaner.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupHFileCleaner.java @@ -30,7 +30,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -77,7 +77,7 @@ public class BackupHFileCleaner extends BaseHFileCleanerDelegate implements Abor if (connection == null) { connection = ConnectionFactory.createConnection(conf); } - try (BackupSystemTable tbl = new BackupSystemTable(connection)) { + try (BackupMetaTable tbl = new BackupMetaTable(connection)) { Map>[] res = tbl.readBulkLoadedFiles(null, tableList); secondPrevReadFromBackupTbl = prevReadFromBackupTbl; @@ -99,7 +99,7 @@ public class BackupHFileCleaner extends BaseHFileCleanerDelegate implements Abor // so that we filter BulkLoad to be returned from server if (checkForFullyBackedUpTables) { if (connection == null) return files; - try (BackupSystemTable tbl = new BackupSystemTable(connection)) { + try (BackupMetaTable tbl = new BackupMetaTable(connection)) { fullyBackedUpTables = tbl.getTablesForBackupType(BackupType.FULL); } catch (IOException ioe) { LOG.error("Failed to get tables which have been fully backed up, skipping checking", ioe); diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupObserver.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupObserver.java index 191e5025dd..c2cbbc7ea7 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupObserver.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/BackupObserver.java @@ -28,7 +28,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.impl.BackupManager; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.RegionInfo; @@ -67,7 +67,7 @@ public class BackupObserver implements RegionCoprocessor, RegionObserver { return; } try (Connection connection = ConnectionFactory.createConnection(cfg); - BackupSystemTable tbl = new BackupSystemTable(connection)) { + BackupMetaTable tbl = new BackupMetaTable(connection)) { List fullyBackedUpTables = tbl.getTablesForBackupType(BackupType.FULL); RegionInfo info = ctx.getEnvironment().getRegionInfo(); TableName tableName = info.getTable(); @@ -91,7 +91,7 @@ public class BackupObserver implements RegionCoprocessor, RegionObserver { return; } try (Connection connection = ConnectionFactory.createConnection(cfg); - BackupSystemTable tbl = new BackupSystemTable(connection)) { + BackupMetaTable tbl = new BackupMetaTable(connection)) { List fullyBackedUpTables = tbl.getTablesForBackupType(BackupType.FULL); RegionInfo info = ctx.getEnvironment().getRegionInfo(); TableName tableName = info.getTable(); diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java index 2903bb0d2a..796a1d7f1c 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java @@ -45,7 +45,7 @@ import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.impl.BackupAdminImpl; import org.apache.hadoop.hbase.backup.impl.BackupManager; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.util.BackupUtils; import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; @@ -188,7 +188,7 @@ public class RestoreDriver extends AbstractHBaseTool { private String getTablesForSet(Connection conn, String name, Configuration conf) throws IOException { - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { List tables = table.describeBackupSet(name); if (tables == null) return null; return StringUtils.join(tables, BackupRestoreConstants.TABLENAME_DELIMITER_IN_COMMAND); diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.java index f27490c907..bd8828e803 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.java @@ -72,7 +72,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public BackupInfo getBackupInfo(String backupId) throws IOException { BackupInfo backupInfo = null; - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { if (backupId == null) { ArrayList recentSessions = table.getBackupInfos(BackupState.RUNNING); if (recentSessions.isEmpty()) { @@ -97,7 +97,7 @@ public class BackupAdminImpl implements BackupAdmin { boolean deleteSessionStarted = false; boolean snapshotDone = false; - try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + try (final BackupMetaTable sysTable = new BackupMetaTable(conn)) { // Step 1: Make sure there is no active session // is running by using startBackupSession API @@ -123,8 +123,8 @@ public class BackupAdminImpl implements BackupAdmin { // Step 3: Record delete session sysTable.startDeleteOperation(backupIds); // Step 4: Snapshot backup system table - if (!BackupSystemTable.snapshotExists(conn)) { - BackupSystemTable.snapshot(conn); + if (!BackupMetaTable.snapshotExists(conn)) { + BackupMetaTable.snapshot(conn); } else { LOG.warn("Backup system table snapshot exists"); } @@ -147,15 +147,15 @@ public class BackupAdminImpl implements BackupAdmin { // Finish sysTable.finishDeleteOperation(); // delete snapshot - BackupSystemTable.deleteSnapshot(conn); + BackupMetaTable.deleteSnapshot(conn); } catch (IOException e) { // Fail delete operation // Step 1 if (snapshotDone) { - if (BackupSystemTable.snapshotExists(conn)) { - BackupSystemTable.restoreFromSnapshot(conn); + if (BackupMetaTable.snapshotExists(conn)) { + BackupMetaTable.restoreFromSnapshot(conn); // delete snapshot - BackupSystemTable.deleteSnapshot(conn); + BackupMetaTable.deleteSnapshot(conn); // We still have record with unfinished delete operation LOG.error("Delete operation failed, please run backup repair utility to restore " + "backup system integrity", e); @@ -181,7 +181,7 @@ public class BackupAdminImpl implements BackupAdmin { * @throws IOException */ - private void finalizeDelete(Map> tablesMap, BackupSystemTable table) + private void finalizeDelete(Map> tablesMap, BackupMetaTable table) throws IOException { for (String backupRoot : tablesMap.keySet()) { Set incrTableSet = table.getIncrementalBackupTableSet(backupRoot); @@ -220,7 +220,7 @@ public class BackupAdminImpl implements BackupAdmin { * @return total number of deleted backup images * @throws IOException */ - private int deleteBackup(String backupId, BackupSystemTable sysTable) throws IOException { + private int deleteBackup(String backupId, BackupMetaTable sysTable) throws IOException { BackupInfo backupInfo = sysTable.readBackupInfo(backupId); @@ -284,7 +284,7 @@ public class BackupAdminImpl implements BackupAdmin { } private void - removeTableFromBackupImage(BackupInfo info, TableName tn, BackupSystemTable sysTable) + removeTableFromBackupImage(BackupInfo info, TableName tn, BackupMetaTable sysTable) throws IOException { List tables = info.getTableNames(); LOG.debug("Remove " + tn + " from " + info.getBackupId() + " tables=" @@ -308,7 +308,7 @@ public class BackupAdminImpl implements BackupAdmin { } private List getAffectedBackupSessions(BackupInfo backupInfo, TableName tn, - BackupSystemTable table) throws IOException { + BackupMetaTable table) throws IOException { LOG.debug("GetAffectedBackupInfos for: " + backupInfo.getBackupId() + " table=" + tn); long ts = backupInfo.getStartTs(); List list = new ArrayList(); @@ -367,7 +367,7 @@ public class BackupAdminImpl implements BackupAdmin { } } - private boolean isLastBackupSession(BackupSystemTable table, TableName tn, long startTime) + private boolean isLastBackupSession(BackupMetaTable table, TableName tn, long startTime) throws IOException { List history = table.getBackupHistory(); for (BackupInfo info : history) { @@ -386,7 +386,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public List getHistory(int n) throws IOException { - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { List history = table.getBackupHistory(); if (history.size() <= n) return history; List list = new ArrayList(); @@ -400,7 +400,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public List getHistory(int n, BackupInfo.Filter... filters) throws IOException { if (filters.length == 0) return getHistory(n); - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { List history = table.getBackupHistory(); List result = new ArrayList(); for (BackupInfo bi : history) { @@ -422,7 +422,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public List listBackupSets() throws IOException { - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { List list = table.listBackupSets(); List bslist = new ArrayList(); for (String s : list) { @@ -437,7 +437,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public BackupSet getBackupSet(String name) throws IOException { - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { List list = table.describeBackupSet(name); if (list == null) return null; return new BackupSet(name, list); @@ -446,7 +446,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public boolean deleteBackupSet(String name) throws IOException { - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { if (table.describeBackupSet(name) == null) { return false; } @@ -458,7 +458,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public void addToBackupSet(String name, TableName[] tables) throws IOException { String[] tableNames = new String[tables.length]; - try (final BackupSystemTable table = new BackupSystemTable(conn); + try (final BackupMetaTable table = new BackupMetaTable(conn); final Admin admin = conn.getAdmin()) { for (int i = 0; i < tables.length; i++) { tableNames[i] = tables[i].getNameAsString(); @@ -475,7 +475,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public void removeFromBackupSet(String name, TableName[] tables) throws IOException { LOG.info("Removing tables [" + StringUtils.join(tables, " ") + "] from '" + name + "'"); - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { table.removeFromBackupSet(name, toStringArray(tables)); LOG.info("Removing tables [" + StringUtils.join(tables, " ") + "] from '" + name + "' completed."); @@ -523,7 +523,7 @@ public class BackupAdminImpl implements BackupAdmin { String backupId = BackupRestoreConstants.BACKUPID_PREFIX + EnvironmentEdgeManager.currentTime(); if (type == BackupType.INCREMENTAL) { Set incrTableSet = null; - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { incrTableSet = table.getIncrementalBackupTableSet(targetRootDir); } @@ -615,7 +615,7 @@ public class BackupAdminImpl implements BackupAdmin { @Override public void mergeBackups(String[] backupIds) throws IOException { - try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + try (final BackupMetaTable sysTable = new BackupMetaTable(conn)) { checkIfValidForMerge(backupIds, sysTable); BackupMergeJob job = BackupRestoreFactory.getBackupMergeJob(conn.getConfiguration()); job.run(backupIds); @@ -636,7 +636,7 @@ public class BackupAdminImpl implements BackupAdmin { * @param table backup system table * @throws IOException */ - private void checkIfValidForMerge(String[] backupIds, BackupSystemTable table) throws IOException { + private void checkIfValidForMerge(String[] backupIds, BackupMetaTable table) throws IOException { String backupRoot = null; final Set allTables = new HashSet(); diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupCommands.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupCommands.java index bbedcee813..2b9c6d7407 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupCommands.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupCommands.java @@ -144,7 +144,7 @@ public final class BackupCommands { conn = ConnectionFactory.createConnection(getConf()); if (requiresNoActiveSession()) { // Check active session - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { List sessions = table.getBackupInfos(BackupState.RUNNING); if (sessions.size() > 0) { @@ -158,7 +158,7 @@ public final class BackupCommands { } if (requiresConsistentState()) { // Check failed delete - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] ids = table.getListOfBackupIdsFromDeleteOperation(); if (ids != null && ids.length > 0) { @@ -368,7 +368,7 @@ public final class BackupCommands { } private String getTablesForSet(String name, Configuration conf) throws IOException { - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { List tables = table.describeBackupSet(name); if (tables == null) return null; return StringUtils.join(tables, BackupRestoreConstants.TABLENAME_DELIMITER_IN_COMMAND); @@ -470,7 +470,7 @@ public final class BackupCommands { super.execute(); String backupId = args[1]; - try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + try (final BackupMetaTable sysTable = new BackupMetaTable(conn)) { BackupInfo info = sysTable.readBackupInfo(backupId); if (info == null) { System.out.println("ERROR: " + backupId + " does not exist"); @@ -511,7 +511,7 @@ public final class BackupCommands { super.execute(); String backupId = (args == null || args.length <= 1) ? null : args[1]; - try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + try (final BackupMetaTable sysTable = new BackupMetaTable(conn)) { BackupInfo info = null; if (backupId != null) { @@ -605,7 +605,7 @@ public final class BackupCommands { Configuration conf = getConf() != null ? getConf() : HBaseConfiguration.create(); try (final Connection conn = ConnectionFactory.createConnection(conf); - final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + final BackupMetaTable sysTable = new BackupMetaTable(conn)) { // Failed backup BackupInfo backupInfo; List list = sysTable.getBackupInfos(BackupState.RUNNING); @@ -641,19 +641,19 @@ public final class BackupCommands { } } - private void repairFailedBackupDeletionIfAny(Connection conn, BackupSystemTable sysTable) + private void repairFailedBackupDeletionIfAny(Connection conn, BackupMetaTable sysTable) throws IOException { String[] backupIds = sysTable.getListOfBackupIdsFromDeleteOperation(); if (backupIds == null || backupIds.length == 0) { System.out.println("No failed backup DELETE operation found"); // Delete backup table snapshot if exists - BackupSystemTable.deleteSnapshot(conn); + BackupMetaTable.deleteSnapshot(conn); return; } System.out.println("Found failed DELETE operation for: " + StringUtils.join(backupIds)); System.out.println("Running DELETE again ..."); // Restore table from snapshot - BackupSystemTable.restoreFromSnapshot(conn); + BackupMetaTable.restoreFromSnapshot(conn); // Finish previous failed session sysTable.finishBackupExclusiveOperation(); try (BackupAdmin admin = new BackupAdminImpl(conn)) { @@ -663,19 +663,19 @@ public final class BackupCommands { } - private void repairFailedBackupMergeIfAny(Connection conn, BackupSystemTable sysTable) + private void repairFailedBackupMergeIfAny(Connection conn, BackupMetaTable sysTable) throws IOException { String[] backupIds = sysTable.getListOfBackupIdsFromMergeOperation(); if (backupIds == null || backupIds.length == 0) { System.out.println("No failed backup MERGE operation found"); // Delete backup table snapshot if exists - BackupSystemTable.deleteSnapshot(conn); + BackupMetaTable.deleteSnapshot(conn); return; } System.out.println("Found failed MERGE operation for: " + StringUtils.join(backupIds)); System.out.println("Running MERGE again ..."); // Restore table from snapshot - BackupSystemTable.restoreFromSnapshot(conn); + BackupMetaTable.restoreFromSnapshot(conn); // Unlock backupo system sysTable.finishBackupExclusiveOperation(); // Finish previous failed session @@ -779,7 +779,7 @@ public final class BackupCommands { if (backupRootPath == null) { // Load from backup system table super.execute(); - try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + try (final BackupMetaTable sysTable = new BackupMetaTable(conn)) { history = sysTable.getBackupHistory(n, tableNameFilter, tableSetFilter); } } else { @@ -919,7 +919,7 @@ public final class BackupCommands { super.execute(); String setName = args[2]; - try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { + try (final BackupMetaTable sysTable = new BackupMetaTable(conn)) { List tables = sysTable.describeBackupSet(setName); BackupSet set = tables == null ? null : new BackupSet(setName, tables); if (set == null) { diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupManager.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupManager.java index 4ca998ca72..247fc0e22f 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupManager.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupManager.java @@ -63,7 +63,7 @@ public class BackupManager implements Closeable { protected Configuration conf = null; protected BackupInfo backupInfo = null; - protected BackupSystemTable systemTable; + protected BackupMetaTable systemTable; protected final Connection conn; /** @@ -80,7 +80,7 @@ public class BackupManager implements Closeable { } this.conf = conf; this.conn = conn; - this.systemTable = new BackupSystemTable(conn); + this.systemTable = new BackupMetaTable(conn); } @@ -214,7 +214,7 @@ public class BackupManager implements Closeable { tableList = new ArrayList<>(); for (HTableDescriptor hTableDescriptor : htds) { TableName tn = hTableDescriptor.getTableName(); - if (tn.equals(BackupSystemTable.getTableName(conf))) { + if (tn.equals(BackupMetaTable.getTableName(conf))) { // skip backup system table continue; } @@ -493,7 +493,7 @@ public class BackupManager implements Closeable { * @return WAL files iterator from backup system table * @throws IOException */ - public Iterator getWALFilesFromBackupSystem() throws IOException { + public Iterator getWALFilesFromBackupSystem() throws IOException { return systemTable.getWALFilesIterator(backupInfo.getBackupRootDir()); } diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupMetaTable.java similarity index 92% rename from hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java rename to hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupMetaTable.java index cf34d14d25..796f720318 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupMetaTable.java @@ -75,9 +75,9 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos; import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; /** - * This class provides API to access backup system table
+ * This class provides API to access backup meta table
* - * Backup system table schema:
+ * Backup meta table schema:
*

    *
  • 1. Backup sessions rowkey= "session:"+backupId; value =serialized BackupInfo
  • *
  • 2. Backup start code rowkey = "startcode:"+backupRoot; value = startcode
  • @@ -91,8 +91,8 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; */ @InterfaceAudience.Private -public final class BackupSystemTable implements Closeable { - private static final Logger LOG = LoggerFactory.getLogger(BackupSystemTable.class); +public final class BackupMetaTable implements Closeable { + private static final Logger LOG = LoggerFactory.getLogger(BackupMetaTable.class); static class WALItem { String backupId; @@ -188,11 +188,11 @@ public final class BackupSystemTable implements Closeable { // Safe delimiter in a string private final static String NULL = "\u0000"; - public BackupSystemTable(Connection conn) throws IOException { + public BackupMetaTable(Connection conn) throws IOException { this.connection = conn; Configuration conf = this.connection.getConfiguration(); - tableName = BackupSystemTable.getTableName(conf); - bulkLoadTableName = BackupSystemTable.getTableNameForBulkLoadedData(conf); + tableName = BackupMetaTable.getTableName(conf); + bulkLoadTableName = BackupMetaTable.getTableNameForBulkLoadedData(conf); checkSystemTable(); } @@ -202,12 +202,12 @@ public final class BackupSystemTable implements Closeable { Configuration conf = connection.getConfiguration(); if (!admin.tableExists(tableName)) { TableDescriptor backupHTD = - BackupSystemTable.getSystemTableDescriptor(conf); + BackupMetaTable.getSystemTableDescriptor(conf); admin.createTable(backupHTD); } if (!admin.tableExists(bulkLoadTableName)) { TableDescriptor blHTD = - BackupSystemTable.getSystemTableForBulkLoadedDataDescriptor(conf); + BackupMetaTable.getSystemTableForBulkLoadedDataDescriptor(conf); admin.createTable(blHTD); } waitForSystemTable(admin, tableName); @@ -276,7 +276,7 @@ public final class BackupSystemTable implements Closeable { * @return Map of rows to path of bulk loaded hfile */ Map readBulkLoadedFiles(String backupId) throws IOException { - Scan scan = BackupSystemTable.createScanForBulkLoadedFiles(backupId); + Scan scan = BackupMetaTable.createScanForBulkLoadedFiles(backupId); try (Table table = connection.getTable(bulkLoadTableName); ResultScanner scanner = table.getScanner(scan)) { Result res = null; @@ -285,8 +285,8 @@ public final class BackupSystemTable implements Closeable { res.advance(); byte[] row = CellUtil.cloneRow(res.listCells().get(0)); for (Cell cell : res.listCells()) { - if (CellUtil.compareQualifiers(cell, BackupSystemTable.PATH_COL, 0, - BackupSystemTable.PATH_COL.length) == 0) { + if (CellUtil.compareQualifiers(cell, BackupMetaTable.PATH_COL, 0, + BackupMetaTable.PATH_COL.length) == 0) { map.put(row, Bytes.toString(CellUtil.cloneValue(cell))); } } @@ -303,7 +303,7 @@ public final class BackupSystemTable implements Closeable { */ public Map>[] readBulkLoadedFiles(String backupId, List sTableList) throws IOException { - Scan scan = BackupSystemTable.createScanForBulkLoadedFiles(backupId); + Scan scan = BackupMetaTable.createScanForBulkLoadedFiles(backupId); Map>[] mapForSrc = new Map[sTableList == null ? 1 : sTableList.size()]; try (Table table = connection.getTable(bulkLoadTableName); ResultScanner scanner = table.getScanner(scan)) { @@ -314,14 +314,14 @@ public final class BackupSystemTable implements Closeable { byte[] fam = null; String path = null; for (Cell cell : res.listCells()) { - if (CellUtil.compareQualifiers(cell, BackupSystemTable.TBL_COL, 0, - BackupSystemTable.TBL_COL.length) == 0) { + if (CellUtil.compareQualifiers(cell, BackupMetaTable.TBL_COL, 0, + BackupMetaTable.TBL_COL.length) == 0) { tbl = TableName.valueOf(CellUtil.cloneValue(cell)); - } else if (CellUtil.compareQualifiers(cell, BackupSystemTable.FAM_COL, 0, - BackupSystemTable.FAM_COL.length) == 0) { + } else if (CellUtil.compareQualifiers(cell, BackupMetaTable.FAM_COL, 0, + BackupMetaTable.FAM_COL.length) == 0) { fam = CellUtil.cloneValue(cell); - } else if (CellUtil.compareQualifiers(cell, BackupSystemTable.PATH_COL, 0, - BackupSystemTable.PATH_COL.length) == 0) { + } else if (CellUtil.compareQualifiers(cell, BackupMetaTable.PATH_COL, 0, + BackupMetaTable.PATH_COL.length) == 0) { path = Bytes.toString(CellUtil.cloneValue(cell)); } } @@ -381,7 +381,7 @@ public final class BackupSystemTable implements Closeable { + " entries"); } try (Table table = connection.getTable(bulkLoadTableName)) { - List puts = BackupSystemTable.createPutForCommittedBulkload(tabName, region, finalPaths); + List puts = BackupMetaTable.createPutForCommittedBulkload(tabName, region, finalPaths); table.put(puts); LOG.debug("written " + puts.size() + " rows for bulk load of " + tabName); } @@ -402,7 +402,7 @@ public final class BackupSystemTable implements Closeable { } try (Table table = connection.getTable(bulkLoadTableName)) { List puts = - BackupSystemTable.createPutForPreparedBulkload(tabName, region, family, pairs); + BackupMetaTable.createPutForPreparedBulkload(tabName, region, family, pairs); table.put(puts); LOG.debug("written " + puts.size() + " rows for bulk load of " + tabName); } @@ -437,7 +437,7 @@ public final class BackupSystemTable implements Closeable { Map>>>> map = new HashMap<>(); List rows = new ArrayList<>(); for (TableName tTable : tableList) { - Scan scan = BackupSystemTable.createScanForOrigBulkLoadedFiles(tTable); + Scan scan = BackupMetaTable.createScanForOrigBulkLoadedFiles(tTable); Map>>> tblMap = map.get(tTable); try (Table table = connection.getTable(bulkLoadTableName); ResultScanner scanner = table.getScanner(scan)) { @@ -453,17 +453,17 @@ public final class BackupSystemTable implements Closeable { row = CellUtil.cloneRow(cell); rows.add(row); String rowStr = Bytes.toString(row); - region = BackupSystemTable.getRegionNameFromOrigBulkLoadRow(rowStr); - if (CellUtil.compareQualifiers(cell, BackupSystemTable.FAM_COL, 0, - BackupSystemTable.FAM_COL.length) == 0) { + region = BackupMetaTable.getRegionNameFromOrigBulkLoadRow(rowStr); + if (CellUtil.compareQualifiers(cell, BackupMetaTable.FAM_COL, 0, + BackupMetaTable.FAM_COL.length) == 0) { fam = Bytes.toString(CellUtil.cloneValue(cell)); - } else if (CellUtil.compareQualifiers(cell, BackupSystemTable.PATH_COL, 0, - BackupSystemTable.PATH_COL.length) == 0) { + } else if (CellUtil.compareQualifiers(cell, BackupMetaTable.PATH_COL, 0, + BackupMetaTable.PATH_COL.length) == 0) { path = Bytes.toString(CellUtil.cloneValue(cell)); - } else if (CellUtil.compareQualifiers(cell, BackupSystemTable.STATE_COL, 0, - BackupSystemTable.STATE_COL.length) == 0) { + } else if (CellUtil.compareQualifiers(cell, BackupMetaTable.STATE_COL, 0, + BackupMetaTable.STATE_COL.length) == 0) { byte[] state = CellUtil.cloneValue(cell); - if (Bytes.equals(BackupSystemTable.BL_PREPARE, state)) { + if (Bytes.equals(BackupMetaTable.BL_PREPARE, state)) { raw = true; } else raw = false; } @@ -507,7 +507,7 @@ public final class BackupSystemTable implements Closeable { List paths = entry.getValue(); for (Path p : paths) { Put put = - BackupSystemTable.createPutForBulkLoadedFile(tn, fam, p.toString(), backupId, ts, + BackupMetaTable.createPutForBulkLoadedFile(tn, fam, p.toString(), backupId, ts, cnt++); puts.add(put); } @@ -1404,7 +1404,7 @@ public final class BackupSystemTable implements Closeable { */ private Put createPutForBackupInfo(BackupInfo context) throws IOException { Put put = new Put(rowkey(BACKUP_INFO_PREFIX, context.getBackupId())); - put.addColumn(BackupSystemTable.SESSIONS_FAMILY, Bytes.toBytes("context"), + put.addColumn(BackupMetaTable.SESSIONS_FAMILY, Bytes.toBytes("context"), context.toByteArray()); return put; } @@ -1417,7 +1417,7 @@ public final class BackupSystemTable implements Closeable { */ private Get createGetForBackupInfo(String backupId) throws IOException { Get get = new Get(rowkey(BACKUP_INFO_PREFIX, backupId)); - get.addFamily(BackupSystemTable.SESSIONS_FAMILY); + get.addFamily(BackupMetaTable.SESSIONS_FAMILY); get.setMaxVersions(1); return get; } @@ -1430,7 +1430,7 @@ public final class BackupSystemTable implements Closeable { */ private Delete createDeleteForBackupInfo(String backupId) { Delete del = new Delete(rowkey(BACKUP_INFO_PREFIX, backupId)); - del.addFamily(BackupSystemTable.SESSIONS_FAMILY); + del.addFamily(BackupMetaTable.SESSIONS_FAMILY); return del; } @@ -1453,7 +1453,7 @@ public final class BackupSystemTable implements Closeable { */ private Get createGetForStartCode(String rootPath) throws IOException { Get get = new Get(rowkey(START_CODE_ROW, rootPath)); - get.addFamily(BackupSystemTable.META_FAMILY); + get.addFamily(BackupMetaTable.META_FAMILY); get.setMaxVersions(1); return get; } @@ -1465,7 +1465,7 @@ public final class BackupSystemTable implements Closeable { */ private Put createPutForStartCode(String startCode, String rootPath) { Put put = new Put(rowkey(START_CODE_ROW, rootPath)); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("startcode"), + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("startcode"), Bytes.toBytes(startCode)); return put; } @@ -1477,7 +1477,7 @@ public final class BackupSystemTable implements Closeable { */ private Get createGetForIncrBackupTableSet(String backupRoot) throws IOException { Get get = new Get(rowkey(INCR_BACKUP_SET, backupRoot)); - get.addFamily(BackupSystemTable.META_FAMILY); + get.addFamily(BackupMetaTable.META_FAMILY); get.setMaxVersions(1); return get; } @@ -1490,7 +1490,7 @@ public final class BackupSystemTable implements Closeable { private Put createPutForIncrBackupTableSet(Set tables, String backupRoot) { Put put = new Put(rowkey(INCR_BACKUP_SET, backupRoot)); for (TableName table : tables) { - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes(table.getNameAsString()), + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes(table.getNameAsString()), EMPTY_VALUE); } return put; @@ -1503,7 +1503,7 @@ public final class BackupSystemTable implements Closeable { */ private Delete createDeleteForIncrBackupTableSet(String backupRoot) { Delete delete = new Delete(rowkey(INCR_BACKUP_SET, backupRoot)); - delete.addFamily(BackupSystemTable.META_FAMILY); + delete.addFamily(BackupMetaTable.META_FAMILY); return delete; } @@ -1518,7 +1518,7 @@ public final class BackupSystemTable implements Closeable { stopRow[stopRow.length - 1] = (byte) (stopRow[stopRow.length - 1] + 1); scan.setStartRow(startRow); scan.setStopRow(stopRow); - scan.addFamily(BackupSystemTable.SESSIONS_FAMILY); + scan.addFamily(BackupMetaTable.SESSIONS_FAMILY); scan.setMaxVersions(1); return scan; } @@ -1543,7 +1543,7 @@ public final class BackupSystemTable implements Closeable { private Put createPutForWriteRegionServerLogTimestamp(TableName table, byte[] smap, String backupRoot) { Put put = new Put(rowkey(TABLE_RS_LOG_MAP_PREFIX, backupRoot, NULL, table.getNameAsString())); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("log-roll-map"), smap); + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("log-roll-map"), smap); return put; } @@ -1558,7 +1558,7 @@ public final class BackupSystemTable implements Closeable { stopRow[stopRow.length - 1] = (byte) (stopRow[stopRow.length - 1] + 1); scan.setStartRow(startRow); scan.setStopRow(stopRow); - scan.addFamily(BackupSystemTable.META_FAMILY); + scan.addFamily(BackupMetaTable.META_FAMILY); return scan; } @@ -1583,7 +1583,7 @@ public final class BackupSystemTable implements Closeable { private Put createPutForRegionServerLastLogRollResult(String server, Long timestamp, String backupRoot) { Put put = new Put(rowkey(RS_LOG_TS_PREFIX, backupRoot, NULL, server)); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("rs-log-ts"), + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("rs-log-ts"), Bytes.toBytes(timestamp)); return put; } @@ -1599,7 +1599,7 @@ public final class BackupSystemTable implements Closeable { stopRow[stopRow.length - 1] = (byte) (stopRow[stopRow.length - 1] + 1); scan.setStartRow(startRow); scan.setStopRow(stopRow); - scan.addFamily(BackupSystemTable.META_FAMILY); + scan.addFamily(BackupMetaTable.META_FAMILY); scan.setMaxVersions(1); return scan; @@ -1630,10 +1630,10 @@ public final class BackupSystemTable implements Closeable { Put put = new Put(rowkey(BULK_LOAD_PREFIX, table.toString(), BLK_LD_DELIM, Bytes.toString(region), BLK_LD_DELIM, filename)); - put.addColumn(BackupSystemTable.META_FAMILY, TBL_COL, table.getName()); - put.addColumn(BackupSystemTable.META_FAMILY, FAM_COL, entry.getKey()); - put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, file.getBytes()); - put.addColumn(BackupSystemTable.META_FAMILY, STATE_COL, BL_COMMIT); + put.addColumn(BackupMetaTable.META_FAMILY, TBL_COL, table.getName()); + put.addColumn(BackupMetaTable.META_FAMILY, FAM_COL, entry.getKey()); + put.addColumn(BackupMetaTable.META_FAMILY, PATH_COL, file.getBytes()); + put.addColumn(BackupMetaTable.META_FAMILY, STATE_COL, BL_COMMIT); puts.add(put); LOG.debug("writing done bulk path " + file + " for " + table + " " + Bytes.toString(region)); } @@ -1644,19 +1644,19 @@ public final class BackupSystemTable implements Closeable { public static void snapshot(Connection conn) throws IOException { try (Admin admin = conn.getAdmin()) { Configuration conf = conn.getConfiguration(); - admin.snapshot(BackupSystemTable.getSnapshotName(conf), BackupSystemTable.getTableName(conf)); + admin.snapshot(BackupMetaTable.getSnapshotName(conf), BackupMetaTable.getTableName(conf)); } } public static void restoreFromSnapshot(Connection conn) throws IOException { Configuration conf = conn.getConfiguration(); - LOG.debug("Restoring " + BackupSystemTable.getTableNameAsString(conf) + " from snapshot"); + LOG.debug("Restoring " + BackupMetaTable.getTableNameAsString(conf) + " from snapshot"); try (Admin admin = conn.getAdmin()) { - String snapshotName = BackupSystemTable.getSnapshotName(conf); + String snapshotName = BackupMetaTable.getSnapshotName(conf); if (snapshotExists(admin, snapshotName)) { - admin.disableTable(BackupSystemTable.getTableName(conf)); + admin.disableTable(BackupMetaTable.getTableName(conf)); admin.restoreSnapshot(snapshotName); - admin.enableTable(BackupSystemTable.getTableName(conf)); + admin.enableTable(BackupMetaTable.getTableName(conf)); LOG.debug("Done restoring backup system table"); } else { // Snapshot does not exists, i.e completeBackup failed after @@ -1684,9 +1684,9 @@ public final class BackupSystemTable implements Closeable { public static void deleteSnapshot(Connection conn) throws IOException { Configuration conf = conn.getConfiguration(); - LOG.debug("Deleting " + BackupSystemTable.getSnapshotName(conf) + " from the system"); + LOG.debug("Deleting " + BackupMetaTable.getSnapshotName(conf) + " from the system"); try (Admin admin = conn.getAdmin()) { - String snapshotName = BackupSystemTable.getSnapshotName(conf); + String snapshotName = BackupMetaTable.getSnapshotName(conf); if (snapshotExists(admin, snapshotName)) { admin.deleteSnapshot(snapshotName); LOG.debug("Done deleting backup system table snapshot"); @@ -1710,10 +1710,10 @@ public final class BackupSystemTable implements Closeable { Put put = new Put(rowkey(BULK_LOAD_PREFIX, table.toString(), BLK_LD_DELIM, Bytes.toString(region), BLK_LD_DELIM, filename)); - put.addColumn(BackupSystemTable.META_FAMILY, TBL_COL, table.getName()); - put.addColumn(BackupSystemTable.META_FAMILY, FAM_COL, family); - put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, file.getBytes()); - put.addColumn(BackupSystemTable.META_FAMILY, STATE_COL, BL_PREPARE); + put.addColumn(BackupMetaTable.META_FAMILY, TBL_COL, table.getName()); + put.addColumn(BackupMetaTable.META_FAMILY, FAM_COL, family); + put.addColumn(BackupMetaTable.META_FAMILY, PATH_COL, file.getBytes()); + put.addColumn(BackupMetaTable.META_FAMILY, STATE_COL, BL_PREPARE); puts.add(put); LOG.debug("writing raw bulk path " + file + " for " + table + " " + Bytes.toString(region)); } @@ -1724,7 +1724,7 @@ public final class BackupSystemTable implements Closeable { List lstDels = new ArrayList<>(lst.size()); for (TableName table : lst) { Delete del = new Delete(rowkey(BULK_LOAD_PREFIX, table.toString(), BLK_LD_DELIM)); - del.addFamily(BackupSystemTable.META_FAMILY); + del.addFamily(BackupMetaTable.META_FAMILY); lstDels.add(del); } return lstDels; @@ -1874,7 +1874,7 @@ public final class BackupSystemTable implements Closeable { stopRow[stopRow.length - 1] = (byte) (stopRow[stopRow.length - 1] + 1); scan.withStartRow(startRow); scan.withStopRow(stopRow); - scan.addFamily(BackupSystemTable.META_FAMILY); + scan.addFamily(BackupMetaTable.META_FAMILY); scan.setMaxVersions(1); return scan; } @@ -1911,7 +1911,7 @@ public final class BackupSystemTable implements Closeable { scan.setStartRow(startRow); scan.setStopRow(stopRow); // scan.setTimeRange(lower, Long.MAX_VALUE); - scan.addFamily(BackupSystemTable.META_FAMILY); + scan.addFamily(BackupMetaTable.META_FAMILY); scan.setMaxVersions(1); return scan; } @@ -1919,9 +1919,9 @@ public final class BackupSystemTable implements Closeable { static Put createPutForBulkLoadedFile(TableName tn, byte[] fam, String p, String backupId, long ts, int idx) { Put put = new Put(rowkey(BULK_LOAD_PREFIX, backupId + BLK_LD_DELIM + ts + BLK_LD_DELIM + idx)); - put.addColumn(BackupSystemTable.META_FAMILY, TBL_COL, tn.getName()); - put.addColumn(BackupSystemTable.META_FAMILY, FAM_COL, fam); - put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, p.getBytes()); + put.addColumn(BackupMetaTable.META_FAMILY, TBL_COL, tn.getName()); + put.addColumn(BackupMetaTable.META_FAMILY, FAM_COL, fam); + put.addColumn(BackupMetaTable.META_FAMILY, PATH_COL, p.getBytes()); return put; } @@ -1938,10 +1938,10 @@ public final class BackupSystemTable implements Closeable { List puts = new ArrayList(files.size()); for (String file : files) { Put put = new Put(rowkey(WALS_PREFIX, BackupUtils.getUniqueWALFileNamePart(file))); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("backupId"), + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("backupId"), Bytes.toBytes(backupId)); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("file"), Bytes.toBytes(file)); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("root"), Bytes.toBytes(backupRoot)); + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("file"), Bytes.toBytes(file)); + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("root"), Bytes.toBytes(backupRoot)); puts.add(put); } return puts; @@ -1960,7 +1960,7 @@ public final class BackupSystemTable implements Closeable { stopRow[stopRow.length - 1] = (byte) (stopRow[stopRow.length - 1] + 1); scan.setStartRow(startRow); scan.setStopRow(stopRow); - scan.addFamily(BackupSystemTable.META_FAMILY); + scan.addFamily(BackupMetaTable.META_FAMILY); return scan; } @@ -1973,7 +1973,7 @@ public final class BackupSystemTable implements Closeable { private Get createGetForCheckWALFile(String file) throws IOException { Get get = new Get(rowkey(WALS_PREFIX, BackupUtils.getUniqueWALFileNamePart(file))); // add backup root column - get.addFamily(BackupSystemTable.META_FAMILY); + get.addFamily(BackupMetaTable.META_FAMILY); return get; } @@ -1988,7 +1988,7 @@ public final class BackupSystemTable implements Closeable { stopRow[stopRow.length - 1] = (byte) (stopRow[stopRow.length - 1] + 1); scan.setStartRow(startRow); scan.setStopRow(stopRow); - scan.addFamily(BackupSystemTable.META_FAMILY); + scan.addFamily(BackupMetaTable.META_FAMILY); return scan; } @@ -1998,7 +1998,7 @@ public final class BackupSystemTable implements Closeable { */ private Get createGetForBackupSet(String name) { Get get = new Get(rowkey(SET_KEY_PREFIX, name)); - get.addFamily(BackupSystemTable.META_FAMILY); + get.addFamily(BackupMetaTable.META_FAMILY); return get; } @@ -2009,7 +2009,7 @@ public final class BackupSystemTable implements Closeable { */ private Delete createDeleteForBackupSet(String name) { Delete del = new Delete(rowkey(SET_KEY_PREFIX, name)); - del.addFamily(BackupSystemTable.META_FAMILY); + del.addFamily(BackupMetaTable.META_FAMILY); return del; } @@ -2022,7 +2022,7 @@ public final class BackupSystemTable implements Closeable { private Put createPutForBackupSet(String name, String[] tables) { Put put = new Put(rowkey(SET_KEY_PREFIX, name)); byte[] value = convertToByteArray(tables); - put.addColumn(BackupSystemTable.META_FAMILY, Bytes.toBytes("tables"), value); + put.addColumn(BackupMetaTable.META_FAMILY, Bytes.toBytes("tables"), value); return put; } diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.java index 547082387f..b2b9342c7e 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.java @@ -33,7 +33,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.PathFilter; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable.WALItem; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable.WALItem; import org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager; import org.apache.hadoop.hbase.backup.util.BackupUtils; import org.apache.hadoop.hbase.client.Admin; diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/TableBackupClient.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/TableBackupClient.java index ab24cca001..1b75823d4b 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/TableBackupClient.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/TableBackupClient.java @@ -110,7 +110,7 @@ public abstract class TableBackupClient { protected void beginBackup(BackupManager backupManager, BackupInfo backupInfo) throws IOException { - BackupSystemTable.snapshot(conn); + BackupMetaTable.snapshot(conn); backupManager.setBackupInfo(backupInfo); // set the start timestamp of the overall backup long startTs = EnvironmentEdgeManager.currentTime(); @@ -269,8 +269,8 @@ public abstract class TableBackupClient { deleteSnapshots(conn, backupInfo, conf); cleanupExportSnapshotLog(conf); } - BackupSystemTable.restoreFromSnapshot(conn); - BackupSystemTable.deleteSnapshot(conn); + BackupMetaTable.restoreFromSnapshot(conn); + BackupMetaTable.deleteSnapshot(conn); // clean up the uncompleted data at target directory if the ongoing backup has already entered // the copy phase // For incremental backup, DistCp logs will be cleaned with the targetDir. @@ -402,7 +402,7 @@ public abstract class TableBackupClient { } else if (type == BackupType.INCREMENTAL) { cleanupDistCpLog(backupInfo, conf); } - BackupSystemTable.deleteSnapshot(conn); + BackupMetaTable.deleteSnapshot(conn); backupManager.updateBackupInfo(backupInfo); // Finish active session diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.java index b1f17cfe53..158d4e0a74 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.backup.BackupMergeJob; import org.apache.hadoop.hbase.backup.BackupRestoreConstants; import org.apache.hadoop.hbase.backup.HBackupFileSystem; import org.apache.hadoop.hbase.backup.impl.BackupManifest; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.util.BackupUtils; import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; @@ -90,7 +90,7 @@ public class MapReduceBackupMergeJob implements BackupMergeJob { List> processedTableList = new ArrayList>(); boolean finishedTables = false; Connection conn = ConnectionFactory.createConnection(getConf()); - BackupSystemTable table = new BackupSystemTable(conn); + BackupMetaTable table = new BackupMetaTable(conn); FileSystem fs = FileSystem.get(getConf()); try { @@ -223,7 +223,7 @@ public class MapReduceBackupMergeJob implements BackupMergeJob { String backupRoot) throws IOException { // Delete from backup system table - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { for (String backupId : backupIds) { table.deleteBackupInfo(backupId); } @@ -286,7 +286,7 @@ public class MapReduceBackupMergeJob implements BackupMergeJob { Set allSet = new HashSet(); try (Connection conn = ConnectionFactory.createConnection(conf); - BackupSystemTable table = new BackupSystemTable(conn)) { + BackupMetaTable table = new BackupMetaTable(conn)) { for (String backupId : backupIds) { BackupInfo bInfo = table.readBackupInfo(backupId); diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/BackupLogCleaner.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/BackupLogCleaner.java index a8ece39584..5042c35e8b 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/BackupLogCleaner.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/BackupLogCleaner.java @@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.backup.BackupRestoreConstants; import org.apache.hadoop.hbase.backup.impl.BackupManager; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -84,7 +84,7 @@ public class BackupLogCleaner extends BaseLogCleanerDelegate { } List list = new ArrayList(); - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { // If we do not have recorded backup sessions try { if (!table.hasBackupSessions()) { diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/regionserver/LogRollBackupSubprocedure.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/regionserver/LogRollBackupSubprocedure.java index 4353b46b17..081e113c45 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/regionserver/LogRollBackupSubprocedure.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/regionserver/LogRollBackupSubprocedure.java @@ -22,7 +22,7 @@ import java.util.HashMap; import java.util.List; import java.util.concurrent.Callable; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.errorhandling.ForeignException; @@ -100,7 +100,7 @@ public class LogRollBackupSubprocedure extends Subprocedure { + " on " + rss.getServerName()); Connection connection = rss.getConnection(); - try (final BackupSystemTable table = new BackupSystemTable(connection)) { + try (final BackupMetaTable table = new BackupMetaTable(connection)) { // sanity check, good for testing HashMap serverTimestampMap = table.readRegionServerLastLogRollResult(backupRoot); diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java index 2be778421f..1b04d32a34 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java @@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase; import org.apache.hadoop.hbase.backup.BackupInfo.BackupState; import org.apache.hadoop.hbase.backup.impl.BackupAdminImpl; import org.apache.hadoop.hbase.backup.impl.BackupManager; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.impl.FullTableBackupClient; import org.apache.hadoop.hbase.backup.impl.IncrementalBackupManager; import org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient; @@ -472,7 +472,7 @@ public class TestBackupBase { } private BackupInfo getBackupInfo(String backupId) throws IOException { - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { BackupInfo status = table.readBackupInfo(backupId); return status; } diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDelete.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDelete.java index 0dd3de9356..cb3e3f74f4 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDelete.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDelete.java @@ -27,7 +27,7 @@ import java.util.List; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.util.ToolRunner; import org.junit.Test; @@ -54,7 +54,7 @@ public class TestBackupDelete extends TestBackupBase { assertTrue(checkSucceeded(backupId)); LOG.info("backup complete"); String[] backupIds = new String[] { backupId }; - BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection()); + BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection()); BackupInfo info = table.readBackupInfo(backupId); Path path = new Path(info.getBackupRootDir(), backupId); FileSystem fs = FileSystem.get(path.toUri(), conf1); diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteWithFailures.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteWithFailures.java index 9447f28b73..448c3c47d4 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteWithFailures.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteWithFailures.java @@ -30,7 +30,7 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.SnapshotDescription; @@ -144,7 +144,7 @@ public class TestBackupDeleteWithFailures extends TestBackupBase{ assertTrue(checkSucceeded(backupId)); LOG.info("backup complete"); String[] backupIds = new String[] { backupId }; - BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection()); + BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection()); BackupInfo info = table.readBackupInfo(backupId); Path path = new Path(info.getBackupRootDir(), backupId); FileSystem fs = FileSystem.get(path.toUri(), conf1); diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDescribe.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDescribe.java index 2f111e2191..8a36df707c 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDescribe.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDescribe.java @@ -28,7 +28,7 @@ import java.util.List; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.BackupInfo.BackupState; import org.apache.hadoop.hbase.backup.impl.BackupCommands; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.util.ToolRunner; import org.junit.Test; @@ -98,7 +98,7 @@ public class TestBackupDescribe extends TestBackupBase { assertTrue(response.indexOf(backupId) > 0); assertTrue(response.indexOf("COMPLETE") > 0); - BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection()); + BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection()); BackupInfo status = table.readBackupInfo(backupId); String desc = status.getShortDescription(); table.close(); diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java index c2c3e59a3b..4e82279cb6 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java @@ -32,7 +32,7 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.testclassification.MasterTests; @@ -119,7 +119,7 @@ public class TestBackupHFileCleaner { List list = new ArrayList<>(1); list.add(file); try (Connection conn = ConnectionFactory.createConnection(conf); - BackupSystemTable sysTbl = new BackupSystemTable(conn)) { + BackupMetaTable sysTbl = new BackupMetaTable(conn)) { List sTableList = new ArrayList<>(); sTableList.add(tableName); Map>[] maps = new Map[1]; diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupSystemTable.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupMetaTable.java similarity index 94% rename from hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupSystemTable.java rename to hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupMetaTable.java index f5ee268b8d..0d6a33533c 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupSystemTable.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupMetaTable.java @@ -44,7 +44,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.BackupInfo.BackupState; import org.apache.hadoop.hbase.backup.impl.BackupManager; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.testclassification.MediumTests; @@ -59,13 +59,13 @@ import org.junit.experimental.categories.Category; * Test cases for backup system table API */ @Category(MediumTests.class) -public class TestBackupSystemTable { +public class TestBackupMetaTable { private static final HBaseTestingUtility UTIL = new HBaseTestingUtility(); protected static Configuration conf = UTIL.getConfiguration(); protected static MiniHBaseCluster cluster; protected static Connection conn; - protected BackupSystemTable table; + protected BackupMetaTable table; @BeforeClass public static void setUp() throws Exception { @@ -78,7 +78,7 @@ public class TestBackupSystemTable { @Before public void before() throws IOException { - table = new BackupSystemTable(conn); + table = new BackupMetaTable(conn); } @After @@ -116,10 +116,10 @@ public class TestBackupSystemTable { private void cleanBackupTable() throws IOException { Admin admin = UTIL.getHBaseAdmin(); - admin.disableTable(BackupSystemTable.getTableName(conf)); - admin.truncateTable(BackupSystemTable.getTableName(conf), true); - if (admin.isTableDisabled(BackupSystemTable.getTableName(conf))) { - admin.enableTable(BackupSystemTable.getTableName(conf)); + admin.disableTable(BackupMetaTable.getTableName(conf)); + admin.truncateTable(BackupMetaTable.getTableName(conf), true); + if (admin.isTableDisabled(BackupMetaTable.getTableName(conf))) { + admin.enableTable(BackupMetaTable.getTableName(conf)); } } @@ -153,7 +153,7 @@ public class TestBackupSystemTable { @Test public void testBackupDelete() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { int n = 10; List list = createBackupInfoList(n); @@ -226,7 +226,7 @@ public class TestBackupSystemTable { tables2.add(TableName.valueOf("t5")); table.addIncrementalBackupTableSet(tables1, "root"); - BackupSystemTable table = new BackupSystemTable(conn); + BackupMetaTable table = new BackupMetaTable(conn); TreeSet res1 = (TreeSet) table.getIncrementalBackupTableSet("root"); assertTrue(tables1.size() == res1.size()); Iterator desc1 = tables1.descendingIterator(); @@ -366,7 +366,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetAddNotExists() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3" }; String setName = "name"; @@ -384,7 +384,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetAddExists() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3" }; String setName = "name"; @@ -407,7 +407,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetAddExistsIntersects() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3" }; String setName = "name"; @@ -430,7 +430,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetRemoveSomeNotExists() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3", "table4" }; String setName = "name"; @@ -452,7 +452,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetRemove() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3", "table4" }; String setName = "name"; @@ -474,7 +474,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetDelete() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3", "table4" }; String setName = "name"; @@ -489,7 +489,7 @@ public class TestBackupSystemTable { @Test public void testBackupSetList() throws IOException { - try (BackupSystemTable table = new BackupSystemTable(conn)) { + try (BackupMetaTable table = new BackupMetaTable(conn)) { String[] tables = new String[] { "table1", "table2", "table3", "table4" }; String setName1 = "name1"; diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupRepair.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupRepair.java index 4c764767f7..50cfb17212 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupRepair.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupRepair.java @@ -24,7 +24,7 @@ import java.util.List; import java.util.Set; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.impl.TableBackupClient; import org.apache.hadoop.hbase.backup.impl.TableBackupClient.Stage; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -59,7 +59,7 @@ public class TestBackupRepair extends TestBackupBase { public void runBackupAndFailAtStageWithRestore(int stage) throws Exception { conf1.setInt(FullTableBackupClientForTest.BACKUP_TEST_MODE_STAGE, stage); - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { int before = table.getBackupHistory().size(); String[] args = new String[] { "create", "full", BACKUP_ROOT_DIR, "-t", diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackup.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackup.java index d040f0ae80..3296c02ed6 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackup.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackup.java @@ -21,7 +21,7 @@ import static org.junit.Assert.assertTrue; import java.util.List; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.util.ToolRunner; import org.junit.Test; @@ -37,7 +37,7 @@ public class TestFullBackup extends TestBackupBase { @Test public void testFullBackupMultipleCommand() throws Exception { LOG.info("test full backup on a multiple tables with data: command-line"); - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { int before = table.getBackupHistory().size(); String[] args = new String[] { "create", "full", BACKUP_ROOT_DIR, "-t", diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSet.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSet.java index 7705b1df21..151c507a15 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSet.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSet.java @@ -25,7 +25,7 @@ import static org.junit.Assert.assertTrue; import java.util.List; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.util.ToolRunner; @@ -49,7 +49,7 @@ public class TestFullBackupSet extends TestBackupBase { LOG.info("Test full backup, backup set exists"); // Create set - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { String name = "name"; table.addToBackupSet(name, new String[] { table1.getNameAsString() }); List names = table.describeBackupSet(name); diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSetRestoreSet.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSetRestoreSet.java index 3649cb24b0..ef0c765bbe 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSetRestoreSet.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupSetRestoreSet.java @@ -25,7 +25,7 @@ import static org.junit.Assert.assertTrue; import java.util.List; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.util.ToolRunner; @@ -45,7 +45,7 @@ public class TestFullBackupSetRestoreSet extends TestBackupBase { LOG.info("Test full restore set"); // Create set - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { String name = "name"; table.addToBackupSet(name, new String[] { table1.getNameAsString() }); List names = table.describeBackupSet(name); @@ -88,7 +88,7 @@ public class TestFullBackupSetRestoreSet extends TestBackupBase { LOG.info("Test full restore set to same table"); // Create set - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { String name = "name1"; table.addToBackupSet(name, new String[] { table1.getNameAsString() }); List names = table.describeBackupSet(name); diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupWithFailures.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupWithFailures.java index 74b2707d17..6ac9a80f6a 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupWithFailures.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestFullBackupWithFailures.java @@ -24,7 +24,7 @@ import java.util.List; import java.util.Set; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.impl.TableBackupClient; import org.apache.hadoop.hbase.backup.impl.TableBackupClient.Stage; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -54,7 +54,7 @@ public class TestFullBackupWithFailures extends TestBackupBase { public void runBackupAndFailAtStage(int stage) throws Exception { conf1.setInt(FullTableBackupClientForTest.BACKUP_TEST_MODE_STAGE, stage); - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { int before = table.getBackupHistory().size(); String[] args = new String[] { "create", "full", BACKUP_ROOT_DIR, "-t", diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupMergeWithFailures.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupMergeWithFailures.java index 0ec78be81b..76c2dc95cc 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupMergeWithFailures.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupMergeWithFailures.java @@ -31,7 +31,7 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.impl.BackupAdminImpl; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.mapreduce.MapReduceBackupMergeJob; import org.apache.hadoop.hbase.backup.mapreduce.MapReduceHFileSplitterJob; import org.apache.hadoop.hbase.backup.util.BackupUtils; @@ -98,7 +98,7 @@ public class TestIncrementalBackupMergeWithFailures extends TestBackupBase { List> processedTableList = new ArrayList>(); boolean finishedTables = false; Connection conn = ConnectionFactory.createConnection(getConf()); - BackupSystemTable table = new BackupSystemTable(conn); + BackupMetaTable table = new BackupMetaTable(conn); FileSystem fs = FileSystem.get(getConf()); try { @@ -272,7 +272,7 @@ public class TestIncrementalBackupMergeWithFailures extends TestBackupBase { bAdmin.mergeBackups(backups); Assert.fail("Expected IOException"); } catch (IOException e) { - BackupSystemTable table = new BackupSystemTable(conn); + BackupMetaTable table = new BackupMetaTable(conn); if(phase.ordinal() < FailurePhase.PHASE4.ordinal()) { // No need to repair: // Both Merge and backup exclusive operations are finished diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithBulkLoad.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithBulkLoad.java index ed1d010119..0cd3b7865a 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithBulkLoad.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithBulkLoad.java @@ -27,7 +27,7 @@ import java.util.Map; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.impl.BackupAdminImpl; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.util.BackupUtils; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; @@ -142,7 +142,7 @@ public class TestIncrementalBackupWithBulkLoad extends TestBackupBase { request = createBackupRequest(BackupType.FULL, tables, BACKUP_ROOT_DIR); backupIdFull = client.backupTables(request); - try (final BackupSystemTable table = new BackupSystemTable(conn)) { + try (final BackupMetaTable table = new BackupMetaTable(conn)) { Pair>>>>, List> pair = table.readBulkloadRows(tables); assertTrue("map still has " + pair.getSecond().size() + " entries", diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithFailures.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithFailures.java index f9c11b08b3..e8cd88e540 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithFailures.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestIncrementalBackupWithFailures.java @@ -30,7 +30,7 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.BackupInfo.BackupState; import org.apache.hadoop.hbase.backup.impl.BackupAdminImpl; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.backup.impl.TableBackupClient; import org.apache.hadoop.hbase.backup.impl.TableBackupClient.Stage; import org.apache.hadoop.hbase.client.Connection; @@ -136,7 +136,7 @@ public class TestIncrementalBackupWithFailures extends TestBackupBase { private void runBackupAndFailAtStage(int stage) throws Exception { conf1.setInt(FullTableBackupClientForTest.BACKUP_TEST_MODE_STAGE, stage); - try (BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection())) { int before = table.getBackupHistory().size(); String[] args = new String[] { "create", "incremental", BACKUP_ROOT_DIR, "-t", diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.java index bb0052d0d4..f0c8014e94 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.java @@ -25,7 +25,7 @@ import java.util.List; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -49,7 +49,7 @@ public class TestRepairAfterFailedDelete extends TestBackupBase { assertTrue(checkSucceeded(backupId)); LOG.info("backup complete"); String[] backupIds = new String[] { backupId }; - BackupSystemTable table = new BackupSystemTable(TEST_UTIL.getConnection()); + BackupMetaTable table = new BackupMetaTable(TEST_UTIL.getConnection()); BackupInfo info = table.readBackupInfo(backupId); Path path = new Path(info.getBackupRootDir(), backupId); FileSystem fs = FileSystem.get(path.toUri(), conf1); @@ -59,7 +59,7 @@ public class TestRepairAfterFailedDelete extends TestBackupBase { String snapshotName = "snapshot-backup"; Connection conn = TEST_UTIL.getConnection(); Admin admin = conn.getAdmin(); - admin.snapshot(snapshotName, BackupSystemTable.getTableName(conf1)); + admin.snapshot(snapshotName, BackupMetaTable.getTableName(conf1)); int deleted = getBackupAdmin().deleteBackups(backupIds); @@ -69,9 +69,9 @@ public class TestRepairAfterFailedDelete extends TestBackupBase { // Emulate delete failure // Restore backup system table - admin.disableTable(BackupSystemTable.getTableName(conf1)); + admin.disableTable(BackupMetaTable.getTableName(conf1)); admin.restoreSnapshot(snapshotName); - admin.enableTable(BackupSystemTable.getTableName(conf1)); + admin.enableTable(BackupMetaTable.getTableName(conf1)); // Start backup session table.startBackupExclusiveOperation(); // Start delete operation diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.java index 31d11e2da4..156258c0fa 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.backup; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.junit.experimental.categories.Category; @@ -40,7 +40,7 @@ public class TestSystemTableSnapshot extends TestBackupBase { LOG.info("test snapshot system table"); - TableName backupSystem = BackupSystemTable.getTableName(conf1); + TableName backupSystem = BackupMetaTable.getTableName(conf1); HBaseAdmin hba = TEST_UTIL.getHBaseAdmin(); String snapshotName = "sysTable"; diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/master/TestBackupLogCleaner.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/master/TestBackupLogCleaner.java index 3f0a6324fb..fabda4b843 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/master/TestBackupLogCleaner.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/master/TestBackupLogCleaner.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.backup.BackupType; import org.apache.hadoop.hbase.backup.TestBackupBase; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.HTable; @@ -66,7 +66,7 @@ public class TestBackupLogCleaner extends TestBackupBase { List tableSetFullList = Lists.newArrayList(table1, table2, table3, table4); - try (BackupSystemTable systemTable = new BackupSystemTable(TEST_UTIL.getConnection())) { + try (BackupMetaTable systemTable = new BackupMetaTable(TEST_UTIL.getConnection())) { // Verify that we have no backup sessions yet assertFalse(systemTable.hasBackupSessions()); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java index 4c56444aaf..aaaf48e5ac 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestBackupRestore.java @@ -37,7 +37,7 @@ import org.apache.hadoop.hbase.backup.BackupType; import org.apache.hadoop.hbase.backup.RestoreRequest; import org.apache.hadoop.hbase.backup.impl.BackupAdminImpl; import org.apache.hadoop.hbase.backup.impl.BackupManager; -import org.apache.hadoop.hbase.backup.impl.BackupSystemTable; +import org.apache.hadoop.hbase.backup.impl.BackupMetaTable; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.HTable; @@ -228,7 +228,7 @@ public class IntegrationTestBackupRestore extends IntegrationTestBase { } private BackupInfo getBackupInfo(String backupId) throws IOException { - try (BackupSystemTable table = new BackupSystemTable(util.getConnection())) { + try (BackupMetaTable table = new BackupMetaTable(util.getConnection())) { return table.readBackupInfo(backupId); } } -- 2.11.0 (Apple Git-81)