Index: hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/DefaultOutputFormatContainer.java IDEA additional info: Subsystem: com.intellij.openapi.diff.impl.patch.CharsetEP <+>UTF-8 =================================================================== --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/DefaultOutputFormatContainer.java (revision c7e1d34b6e8734ca65e9a635f4094876b2032d77) +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/DefaultOutputFormatContainer.java (revision ) @@ -51,8 +51,9 @@ super(of); } - static synchronized String getOutputName(int partition) { - return "part-" + NUMBER_FORMAT.format(partition); + static synchronized String getOutputName(TaskAttemptContext context) { + return context.getConfiguration().get("mapreduce.output.basename", "part") + + "-" + NUMBER_FORMAT.format(context.getTaskAttemptID().getTaskID().getId()); } /** @@ -65,7 +66,7 @@ @Override public RecordWriter, HCatRecord> getRecordWriter(TaskAttemptContext context) throws IOException, InterruptedException { - String name = getOutputName(context.getTaskAttemptID().getTaskID().getId()); + String name = getOutputName(context); return new DefaultRecordWriterContainer(context, getBaseOutputFormat().getRecordWriter(null, new JobConf(context.getConfiguration()), name, InternalUtil.createReporter(context))); } Index: hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/DynamicPartitionFileRecordWriterContainer.java IDEA additional info: Subsystem: com.intellij.openapi.diff.impl.patch.CharsetEP <+>UTF-8 =================================================================== --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/DynamicPartitionFileRecordWriterContainer.java (revision c7e1d34b6e8734ca65e9a635f4094876b2032d77) +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/DynamicPartitionFileRecordWriterContainer.java (revision ) @@ -205,7 +205,8 @@ Path parentDir = new Path(currTaskContext.getConfiguration().get("mapred.work.output.dir")); Path childPath = - new Path(parentDir, FileOutputFormat.getUniqueFile(currTaskContext, "part", "")); + new Path(parentDir, FileOutputFormat.getUniqueFile(currTaskContext, + currTaskContext.getConfiguration().get("mapreduce.output.basename", "part"), "")); RecordWriter baseRecordWriter = baseOF.getRecordWriter(parentDir.getFileSystem(currTaskContext.getConfiguration()), Index: hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputFormatContainer.java IDEA additional info: Subsystem: com.intellij.openapi.diff.impl.patch.CharsetEP <+>UTF-8 =================================================================== --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputFormatContainer.java (revision c7e1d34b6e8734ca65e9a635f4094876b2032d77) +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FileOutputFormatContainer.java (revision ) @@ -97,7 +97,8 @@ (org.apache.hadoop.mapred.RecordWriter)null, context); } else { Path parentDir = new Path(context.getConfiguration().get("mapred.work.output.dir")); - Path childPath = new Path(parentDir,FileOutputFormat.getUniqueName(new JobConf(context.getConfiguration()), "part")); + Path childPath = new Path(parentDir,FileOutputFormat.getUniqueName(new JobConf(context.getConfiguration()), + context.getConfiguration().get("mapreduce.output.basename", "part"))); rw = new StaticPartitionFileRecordWriterContainer( getBaseOutputFormat().getRecordWriter(