.../hadoop/hbase/regionserver/DefaultStoreFlusher.java | 12 ++++++++++-- .../regionserver/compactions/DefaultCompactor.java | 17 +++++++++++++---- 2 files changed, 23 insertions(+), 6 deletions(-) diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java index 32c17bd..ab51a3c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java @@ -31,7 +31,6 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.monitoring.MonitoredTask; -import org.apache.hadoop.hbase.util.CollectionBackedScanner; import org.apache.hadoop.util.StringUtils; /** @@ -71,10 +70,19 @@ public class DefaultStoreFlusher extends StoreFlusher { writer = store.createWriterInTmp( snapshot.size(), store.getFamily().getCompression(), false, true, true); writer.setTimeRangeTracker(snapshotTimeRangeTracker); + IOException e = null; try { flushed = performFlush(scanner, writer, smallestReadPoint); + } catch (IOException ioe) { + e = ioe; + // throw the exception out + throw ioe; } finally { - finalizeWriter(writer, cacheFlushId, status); + if (e != null) { + writer.close(); + } else { + finalizeWriter(writer, cacheFlushId, status); + } } } } finally { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java index 7a479d8..adf76d9 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/DefaultCompactor.java @@ -26,9 +26,9 @@ import java.util.List; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.ScanType; +import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.regionserver.StoreFileScanner; @@ -54,6 +54,7 @@ public class DefaultCompactor extends Compactor { StoreFile.Writer writer = null; List newFiles = new ArrayList(); + IOException e = null; try { InternalScanner scanner = null; try { @@ -87,11 +88,19 @@ public class DefaultCompactor extends Compactor { scanner.close(); } } + } catch (IOException ioe) { + e = ioe; + // Throw the exception; + throw ioe; } finally { if (writer != null) { - writer.appendMetadata(fd.maxSeqId, request.isMajor()); - writer.close(); - newFiles.add(writer.getPath()); + if (e != null) { + writer.close(); + } else { + writer.appendMetadata(fd.maxSeqId, request.isMajor()); + writer.close(); + newFiles.add(writer.getPath()); + } } } return newFiles;