diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java index 2ff9ad3251..c0910f6618 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java @@ -3361,7 +3361,7 @@ public static String getVertexCounterName(String counter, String vertexName) { // T2) x; // If T is empty and T2 contains 100 rows, the user expects: 0, 100 (2 // rows) - if (isEmptyTable && !skipDummy) { + if (isEmptyTable) { pathsToAdd.add(createDummyFileForEmptyTable(job, work, hiveScratchDir, alias)); } } diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java index 806deb5f31..87e57c04d7 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java @@ -353,7 +353,7 @@ private JobConf cloneJobConf(BaseWork work) throws Exception { MapWork mapWork = (MapWork) work; cloned.setBoolean("mapred.task.is.map", true); List inputPaths = Utilities.getInputPaths(cloned, mapWork, - scratchDir, context, false); + scratchDir, context, true); Utilities.setInputPaths(cloned, inputPaths); Utilities.setMapWork(cloned, mapWork, scratchDir, false); Utilities.createTmpDirs(cloned, mapWork);