Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteText.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteText.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteText.java (working copy) @@ -114,7 +114,7 @@ job.setOutputValueClass(DefaultHCatRecord.class); job.setNumReduceTasks(0); HCatOutputFormat.setOutput(job, OutputJobInfo.create(dbName, - outputTableName, null, serverUri, principalID)); + outputTableName, null)); HCatSchema s = HCatInputFormat.getTableSchema(job); System.err.println("INFO: output schema explicitly set for writing:" + s); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreDemo.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreDemo.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreDemo.java (working copy) @@ -117,7 +117,7 @@ dbName, tableName, null, serverUri, principalID)); // initialize HCatOutputFormat HCatOutputFormat.setOutput(job, OutputJobInfo.create( - dbName, outputTableName, outputPartitionKvps, serverUri, principalID)); + dbName, outputTableName, outputPartitionKvps)); // test with and without specifying schema randomly HCatSchema s = HCatInputFormat.getTableSchema(job); System.err.println("INFO: output schema explicitly set for writing:" + s); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/GroupByAge.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/GroupByAge.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/GroupByAge.java (working copy) @@ -118,7 +118,7 @@ job.setOutputKeyClass(WritableComparable.class); job.setOutputValueClass(DefaultHCatRecord.class); HCatOutputFormat.setOutput(job, OutputJobInfo.create(dbName, - outputTableName, null, serverUri, principalID)); + outputTableName, null)); HCatSchema s = HCatOutputFormat.getTableSchema(job); System.err.println("INFO: output schema explicitly set for writing:" + s); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteRC.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteRC.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteRC.java (working copy) @@ -105,7 +105,7 @@ job.setOutputValueClass(DefaultHCatRecord.class); job.setNumReduceTasks(0); HCatOutputFormat.setOutput(job, OutputJobInfo.create(dbName, - outputTableName, null, serverUri, principalID)); + outputTableName, null)); HCatSchema s = HCatInputFormat.getTableSchema(job); System.err.println("INFO: output schema explicitly set for writing:" + s); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteJson.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteJson.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/WriteJson.java (working copy) @@ -103,7 +103,7 @@ job.setOutputValueClass(DefaultHCatRecord.class); job.setNumReduceTasks(0); HCatOutputFormat.setOutput(job, OutputJobInfo.create(dbName, - outputTableName, null, serverUri, principalID)); + outputTableName, null)); HCatSchema s = HCatInputFormat.getTableSchema(job); System.err.println("INFO: output schema explicitly set for writing:" + s); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreComplex.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreComplex.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreComplex.java (working copy) @@ -106,7 +106,7 @@ dbName, tableName, null, serverUri, principalID)); // initialize HCatOutputFormat HCatOutputFormat.setOutput(job, OutputJobInfo.create( - dbName, outputTableName, outputPartitionKvps, serverUri, principalID)); + dbName, outputTableName, outputPartitionKvps)); HCatSchema s = HCatInputFormat.getTableSchema(job); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/ReadWrite.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/ReadWrite.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/ReadWrite.java (working copy) @@ -95,7 +95,7 @@ job.setOutputKeyClass(Text.class); job.setOutputValueClass(DefaultHCatRecord.class); HCatOutputFormat.setOutput(job, OutputJobInfo.create(dbName, - outputTableName, null, serverUri, principalID)); + outputTableName, null)); HCatSchema s = HCatInputFormat.getTableSchema(job); System.err.println("INFO: output schema explicitly set for writing:" + s); Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreNumbers.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreNumbers.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/StoreNumbers.java (working copy) @@ -181,7 +181,7 @@ dbName, tableName, null, serverUri, principalID)); // initialize HCatOutputFormat HCatOutputFormat.setOutput(job, OutputJobInfo.create( - dbName, outputTableName, outputPartitionKvps, serverUri, principalID)); + dbName, outputTableName, outputPartitionKvps)); // test with and without specifying schema randomly HCatSchema s = HCatInputFormat.getTableSchema(job); if(writeToNonPartPigTable) { Index: src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/HBaseReadWrite.java =================================================================== --- src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/HBaseReadWrite.java (revision 1244282) +++ src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/HBaseReadWrite.java (working copy) @@ -159,7 +159,7 @@ job.setOutputKeyClass(WritableComparable.class); job.setOutputValueClass(DefaultHCatRecord.class); HCatOutputFormat.setOutput(job, OutputJobInfo.create(dbName, - tableName, null, serverUri, principalID)); + tableName, null)); boolean succ = job.waitForCompletion(true); Index: src/test/e2e/hcatalog/tests/pig.conf =================================================================== --- src/test/e2e/hcatalog/tests/pig.conf (revision 1244282) +++ src/test/e2e/hcatalog/tests/pig.conf (working copy) @@ -139,7 +139,6 @@ m map, bb array>) STORED AS INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat' - INPUTDRIVER 'org.apache.hcatalog.pig.drivers.LoadFuncBasedInputDriver' OUTPUTDRIVER 'org.apache.hcatalog.pig.drivers.StoreFuncBasedOutputDriver' TBLPROPERTIES ('hcat.pig.loader'='org.apache.pig.builtin.JsonLoader', 'hcat.pig.storer'='org.apache.pig.builtin.JsonStorage', 'hcat.pig.loader.args'= 's:chararray, i:int, d:double, m:map[chararray], bb:{t:(a:int, b:chararray)}', 'hcat.pig.args.delimiter'=' '); \ @@ -158,11 +157,7 @@ name string, age int, gpa double) -stored as rcfile -TBLPROPERTIES ( - 'hcat.isd'='org.apache.hcatalog.rcfile.RCFileInputDriver', - 'hcat.osd'='org.apache.hcatalog.rcfile.RCFileOutputDriver' -); +stored as rcfile; \ ,'pig' => q\a = load 'all100krc' using org.apache.hcatalog.pig.HCatLoader(); b = foreach a generate name, age; Index: src/test/e2e/hcatalog/tests/hadoop.conf =================================================================== --- src/test/e2e/hcatalog/tests/hadoop.conf (revision 1244282) +++ src/test/e2e/hcatalog/tests/hadoop.conf (working copy) @@ -123,7 +123,6 @@ m map, bb array>) STORED AS INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat' - INPUTDRIVER 'org.apache.hcatalog.pig.drivers.LoadFuncBasedInputDriver' OUTPUTDRIVER 'org.apache.hcatalog.pig.drivers.StoreFuncBasedOutputDriver' TBLPROPERTIES ('hcat.pig.loader'='org.apache.pig.builtin.JsonLoader', 'hcat.pig.storer'='org.apache.pig.builtin.JsonStorage', 'hcat.pig.loader.args'= 's:chararray, i:int, d:double, m:map[chararray], bb:{t:(a:int, b:chararray)}', 'hcat.pig.args.delimiter'=' ');\ ,'hadoop' => q\ @@ -142,11 +141,7 @@ name string, age int, gpa double) -stored as rcfile -TBLPROPERTIES ( - 'hcat.isd'='org.apache.hcatalog.rcfile.RCFileInputDriver', - 'hcat.osd'='org.apache.hcatalog.rcfile.RCFileOutputDriver' -); +stored as rcfile; \, ,'hadoop' => q\ jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc hadoop_write_3