commit 10ce05f06e50f24b0f67734ce39212a59f9d5189 Author: Todd Lipcon Date: Thu Oct 6 21:46:14 2011 -0700 HBASE-4551. Small fixes to compile against hadoop 23 diff --git pom.xml pom.xml index 1875472..8904b53 100644 --- pom.xml +++ pom.xml @@ -1517,14 +1517,16 @@ org.apache.hadoop - hadoop-common-test + hadoop-common ${hadoop.version} + test-jar test org.apache.hadoop - hadoop-hdfs-test + hadoop-hdfs ${hadoop.version} + test-jar test diff --git src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index bc8e601..6a33efd 100644 --- src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -1462,35 +1462,6 @@ public class HBaseTestingUtility { } /** - * Set soft and hard limits in namenode. - * You'll get a NPE if you call before you've started a minidfscluster. - * @param soft Soft limit - * @param hard Hard limit - * @throws NoSuchFieldException - * @throws SecurityException - * @throws IllegalAccessException - * @throws IllegalArgumentException - */ - public void setNameNodeNameSystemLeasePeriod(final int soft, final int hard) - throws SecurityException, NoSuchFieldException, IllegalArgumentException, IllegalAccessException { - // TODO: If 0.20 hadoop do one thing, if 0.21 hadoop do another. - // Not available in 0.20 hdfs. Use reflection to make it happen. - - // private NameNode nameNode; - Field field = this.dfsCluster.getClass().getDeclaredField("nameNode"); - field.setAccessible(true); - NameNode nn = (NameNode)field.get(this.dfsCluster); - field = nn.getClass().getDeclaredField("namesystem"); - field.setAccessible(true); - FSNamesystem namesystem = (FSNamesystem)field.get(nn); - - field = namesystem.getClass().getDeclaredField("leaseManager"); - field.setAccessible(true); - LeaseManager lm = (LeaseManager)field.get(namesystem); - lm.setLeasePeriod(100, 50000); - } - - /** * Set maxRecoveryErrorCount in DFSClient. In 0.20 pre-append its hard-coded to 5 and * makes tests linger. Here is the exception you'll see: *
diff --git src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java
index 6077afb..8d94755 100644
--- src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java
+++ src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java
@@ -100,7 +100,6 @@ public class TestWALObserver {
     conf.setInt("dfs.client.block.recovery.retries", 2);
 
     TEST_UTIL.startMiniCluster(1);
-    TEST_UTIL.setNameNodeNameSystemLeasePeriod(100, 10000);
     Path hbaseRootDir =
       TEST_UTIL.getDFSCluster().getFileSystem().makeQualified(new Path("/hbase"));
     LOG.info("hbase.rootdir=" + hbaseRootDir);
diff --git src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLog.java src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLog.java
index 5a2d379..830c17e 100644
--- src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLog.java
+++ src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLog.java
@@ -46,8 +46,9 @@ import org.apache.hadoop.hbase.Coprocessor;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
 import org.apache.hadoop.hbase.coprocessor.SampleRegionWALObserver;
 import org.apache.hadoop.hdfs.DFSClient;
+import org.apache.hadoop.hdfs.DistributedFileSystem;
 import org.apache.hadoop.hdfs.MiniDFSCluster;
-import org.apache.hadoop.hdfs.protocol.FSConstants.SafeModeAction;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
 import org.apache.hadoop.hdfs.server.datanode.DataNode;
 import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
 import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
@@ -371,7 +372,8 @@ public class TestHLog  {
 
     // Stop the cluster.  (ensure restart since we're sharing MiniDFSCluster)
     try {
-      cluster.getNameNode().setSafeMode(SafeModeAction.SAFEMODE_ENTER);
+      DistributedFileSystem dfs = (DistributedFileSystem) cluster.getFileSystem();
+      dfs.setSafeMode(FSConstants.SafeModeAction.SAFEMODE_ENTER);
       cluster.shutdown();
       try {
         // wal.writer.close() will throw an exception,
diff --git src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLogSplit.java src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLogSplit.java
index f260996..7db8248 100644
--- src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLogSplit.java
+++ src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestHLogSplit.java
@@ -140,8 +140,6 @@ public class TestHLogSplit {
     regions = new ArrayList();
     Collections.addAll(regions, "bbb", "ccc");
     InstrumentedSequenceFileLogWriter.activateFailure = false;
-    // Set the soft lease for hdfs to be down from default of 5 minutes or so.
-    TEST_UTIL.setNameNodeNameSystemLeasePeriod(100, 50000);
   }
 
   @After
diff --git src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
index a33c38b..0b35b82 100644
--- src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
+++ src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
@@ -78,7 +78,6 @@ public class TestWALReplay {
     // The below config supported by 0.20-append and CDH3b2
     conf.setInt("dfs.client.block.recovery.retries", 2);
     TEST_UTIL.startMiniDFSCluster(3);
-    TEST_UTIL.setNameNodeNameSystemLeasePeriod(100, 10000);
     Path hbaseRootDir =
       TEST_UTIL.getDFSCluster().getFileSystem().makeQualified(new Path("/hbase"));
     LOG.info("hbase.rootdir=" + hbaseRootDir);