Index: contrib/src/test/results/clientpositive/dboutput.q.out =================================================================== --- contrib/src/test/results/clientpositive/dboutput.q.out (revision 1038445) +++ contrib/src/test/results/clientpositive/dboutput.q.out (working copy) @@ -143,21 +143,17 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: dboutput('jdbc:derby:../build/test_dboutput_db','','','INSERT INTO app_info (kkey,vvalue) VALUES (?,?)',key,value) - type: int - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: dboutput('jdbc:derby:../build/test_dboutput_db','','','INSERT INTO app_info (kkey,vvalue) VALUES (?,?)',key,value) + type: int + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: contrib/src/test/results/clientpositive/serde_typedbytes4.q.out =================================================================== --- contrib/src/test/results/clientpositive/serde_typedbytes4.q.out (revision 1038445) +++ contrib/src/test/results/clientpositive/serde_typedbytes4.q.out (working copy) @@ -50,42 +50,38 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: - expr: UDFToByte(key) - type: tinyint - expr: value - type: string - outputColumnNames: _col0, _col1 - Transform Operator - command: /bin/cat - output info: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - Select Operator - expressions: + Select Operator + expressions: + expr: UDFToByte(key) + type: tinyint + expr: value + type: string + outputColumnNames: _col0, _col1 + Transform Operator + command: /bin/cat + output info: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: expr: _col0 type: string expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator Index: hbase-handler/src/test/results/hbase_pushdown.q.out =================================================================== --- hbase-handler/src/test/results/hbase_pushdown.q.out (revision 1038445) +++ hbase-handler/src/test/results/hbase_pushdown.q.out (working copy) @@ -43,23 +43,19 @@ predicate: expr: (key = 90) type: boolean - Filter Operator - predicate: - expr: (key = 90) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -101,23 +97,19 @@ predicate: expr: ((key = 90) and (value like '%90%')) type: boolean - Filter Operator - predicate: - expr: ((key = 90) and (value like '%90%')) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -161,23 +153,19 @@ predicate: expr: (((key = 90) and (value like '%90%')) and (key = UDFToInteger(value))) type: boolean - Filter Operator - predicate: - expr: (((key = 90) and (value like '%90%')) and (key = UDFToInteger(value))) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -212,23 +200,19 @@ predicate: expr: (((key = 80) and (key = 90)) and (value like '%90%')) type: boolean - Filter Operator - predicate: - expr: (((key = 80) and (key = 90)) and (value like '%90%')) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -295,23 +279,19 @@ predicate: expr: (CASE WHEN ((key = 90)) THEN (2) ELSE (4) END > 3) type: boolean - Filter Operator - predicate: - expr: (CASE WHEN ((key = 90)) THEN (2) ELSE (4) END > 3) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -348,23 +328,19 @@ predicate: expr: ((key = 80) or (value like '%90%')) type: boolean - Filter Operator - predicate: - expr: ((key = 80) or (value like '%90%')) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -397,23 +373,19 @@ predicate: expr: (key = 90) type: boolean - Filter Operator - predicate: - expr: (key = 90) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: hbase-handler/src/test/results/hbase_queries.q.out =================================================================== --- hbase-handler/src/test/results/hbase_queries.q.out (revision 1038445) +++ hbase-handler/src/test/results/hbase_queries.q.out (working copy) @@ -50,32 +50,28 @@ predicate: expr: ((key % 2) = 0) type: boolean - Filter Operator - predicate: - expr: ((key % 2) = 0) - type: boolean + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: key + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: value - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat - output format: org.apache.hadoop.hive.hbase.HiveHBaseTableOutputFormat - serde: org.apache.hadoop.hive.hbase.HBaseSerDe - name: hbase_table_1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat + output format: org.apache.hadoop.hive.hbase.HiveHBaseTableOutputFormat + serde: org.apache.hadoop.hive.hbase.HBaseSerDe + name: hbase_table_1 PREHOOK: query: FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0 @@ -306,24 +302,20 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + outputColumnNames: _col0 + Reduce Output Operator + key expressions: + expr: _col0 type: int - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 0 + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: 0 y:hbase_table_2 TableScan alias: hbase_table_2 @@ -331,31 +323,27 @@ predicate: expr: (key < 120) type: boolean - Filter Operator - predicate: - expr: (key < 120) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: 1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java (revision 1038445) +++ ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java (working copy) @@ -674,6 +674,24 @@ } } + public void removeParent(Operator parent) { + int parentIndex = parentOperators.indexOf(parent); + assert parentIndex != -1; + if (parentOperators.size() == 1) { + parentOperators = null; + } else { + parentOperators.remove(parentIndex); + } + + int childIndex = parent.getChildOperators().indexOf(this); + assert childIndex != -1; + if (parent.getChildOperators().size() == 1) { + parent.setChildOperators(null); + } else { + parent.getChildOperators().remove(childIndex); + } + } + /** * Replace one parent with another at the same position. Chilren of the new * parent are not updated Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java (revision 1038445) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java (working copy) @@ -66,21 +66,30 @@ private RowResolver toRR = null; /** - * this map contains a expr infos. Each key is a node in the expression tree - * and the information for each node is the value which is used while walking - * the tree by its parent. + * Values the expression sub-trees (predicates) that can be pushed down for + * root expression tree. Since there can be more than one alias in an + * expression tree, this is a map from the alias to predicates. */ private final Map> pushdownPreds; + /** - * Values the expression sub-trees (predicates) that can be pushed down for + * Values the expression sub-trees (predicates) that can not be pushed down for * root expression tree. Since there can be more than one alias in an * expression tree, this is a map from the alias to predicates. */ + private final Map> nonFinalPreds; + + /** + * this map contains a expr infos. Each key is a node in the expression tree + * and the information for each node is the value which is used while walking + * the tree by its parent. + */ private final Map exprInfoMap; private boolean isDeterministic = true; public ExprWalkerInfo() { pushdownPreds = new HashMap>(); + nonFinalPreds = new HashMap>(); exprInfoMap = new HashMap(); } @@ -91,6 +100,7 @@ pushdownPreds = new HashMap>(); exprInfoMap = new HashMap(); + nonFinalPreds = new HashMap>(); } /** @@ -214,6 +224,19 @@ } /** + * Adds the passed list of pushDowns for the alias. + * + * @param alias + * @param pushDowns + */ + public void addPushDowns(String alias, List pushDowns) { + if (pushdownPreds.get(alias) == null) { + pushdownPreds.put(alias, new ArrayList()); + } + pushdownPreds.get(alias).addAll(pushDowns); + } + + /** * Returns the list of pushdown expressions for each alias that appear in the * current operator's RowResolver. The exprs in each list can be combined * using conjunction (AND). @@ -225,6 +248,28 @@ } /** + * Adds the specified expr as a non-final candidate + * + * @param expr + */ + public void addNonFinalCandidate(ExprNodeDesc expr) { + String alias = getAlias(expr); + if (nonFinalPreds.get(alias) == null) { + nonFinalPreds.put(alias, new ArrayList()); + } + nonFinalPreds.get(alias).add(expr.clone()); + } + + /** + * Returns list of non-final candidate predicate for each map. + * + * @return + */ + public Map> getNonFinalCandidates() { + return nonFinalPreds; + } + + /** * Merges the specified pushdown predicates with the current class. * * @param ewi Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java (revision 1038445) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java (working copy) @@ -288,6 +288,8 @@ if (ctx.isCandidate(expr)) { ctx.addFinalCandidate(expr); return; + } else if (!FunctionRegistry.isOpAnd(expr)) { + ctx.addNonFinalCandidate(expr); } if (FunctionRegistry.isOpAnd(expr)) { Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java (revision 1038445) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java (working copy) @@ -43,7 +43,6 @@ import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx; import org.apache.hadoop.hive.ql.metadata.HiveStorageHandler; import org.apache.hadoop.hive.ql.metadata.HiveStoragePredicateHandler; -import org.apache.hadoop.hive.ql.metadata.HiveUtils; import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.parse.OpParseContext; import org.apache.hadoop.hive.ql.parse.RowResolver; @@ -88,6 +87,22 @@ // script operator is a black-box to hive so no optimization here // assuming that nothing can be pushed above the script op // same with LIMIT op + // create a filter with all children predicates + OpWalkerInfo owi = (OpWalkerInfo) procCtx; + ExprWalkerInfo unpushedPreds = mergeChildrenPred(nd, owi, null, false); + return createFilter((Operator)nd, unpushedPreds, owi); + } + + } + + public static class UDTFPPD extends DefaultPPD implements NodeProcessor { + @Override + public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, + Object... nodeOutputs) throws SemanticException { + LOG.info("Processing for " + nd.getName() + "(" + + ((Operator) nd).getIdentifier() + ")"); + //Predicates for UDTF wont be candidates for its children. So, nothing to + //optimize here. See lateral_view_ppd.q for example. return null; } @@ -102,6 +117,10 @@ + ((Operator) nd).getIdentifier() + ")"); OpWalkerInfo owi = (OpWalkerInfo) procCtx; + // The lateral view forward operator has 2 children, a SELECT(*) and + // a SELECT(cols) (for the UDTF operator) The child at index 0 is the + // SELECT(*) because that's the way that the DAG was constructed. We + // only want to get the predicates from the SELECT(*). ExprWalkerInfo childPreds = owi .getPrunedPreds((Operator) nd.getChildren() .get(0)); @@ -146,22 +165,28 @@ OpWalkerInfo owi = (OpWalkerInfo) procCtx; Operator op = (Operator) nd; ExprNodeDesc predicate = (((FilterOperator) nd).getConf()).getPredicate(); - // get pushdown predicates for this operator's predicate - ExprWalkerInfo ewi = ExprWalkerProcFactory.extractPushdownPreds(owi, op, - predicate); - if (!ewi.isDeterministic()) { - /* predicate is not deterministic */ - if (op.getChildren() != null && op.getChildren().size() == 1) { - createFilter(op, owi - .getPrunedPreds((Operator) (op - .getChildren().get(0))), owi); + ExprWalkerInfo ewi = new ExprWalkerInfo(); + // if predicate is sampling pred, the filterop is always a child of + // TableScan, so there is no need to push this predicate. + if (!((FilterOperator)op).getConf().getIsSamplingPred()) { + // get pushdown predicates for this operator's predicate + ewi = ExprWalkerProcFactory.extractPushdownPreds(owi, op, predicate); + if (!ewi.isDeterministic()) { + /* predicate is not deterministic */ + if (op.getChildren() != null && op.getChildren().size() == 1) { + createFilter(op, owi + .getPrunedPreds((Operator) (op + .getChildren().get(0))), owi); + } + return null; } - - return null; + // add this filter for deletion, if it does not have non-final candidates + if (ewi.getNonFinalCandidates().values().isEmpty()) { + owi.addCandidateFilterOp((FilterOperator)op); + } + logExpr(nd, ewi); + owi.putPrunedPreds((Operator) nd, ewi); } - - logExpr(nd, ewi); - owi.putPrunedPreds(op, ewi); // merge it with children predicates mergeWithChildrenPred(op, owi, ewi, null, false); @@ -182,8 +207,12 @@ OpWalkerInfo owi = (OpWalkerInfo) procCtx; Set aliases = getQualifiedAliases((JoinOperator) nd, owi .getRowResolver(nd)); - mergeWithChildrenPred(nd, owi, null, aliases, false); - return null; + boolean hasUnpushedPredicates = mergeWithChildrenPred(nd, owi, null, aliases, false); + if (hasUnpushedPredicates) { + aliases = null; + } + ExprWalkerInfo unpushedPreds = mergeChildrenPred(nd, owi, aliases, false); + return createFilter((Operator)nd, unpushedPreds, owi); } /** @@ -283,7 +312,12 @@ Object... nodeOutputs) throws SemanticException { LOG.info("Processing for " + nd.getName() + "(" + ((Operator) nd).getIdentifier() + ")"); - mergeWithChildrenPred(nd, (OpWalkerInfo) procCtx, null, null, false); + OpWalkerInfo owi = (OpWalkerInfo) procCtx; + boolean hasUnpushedPredicates = mergeWithChildrenPred(nd, owi, null, null, false); + if (hasUnpushedPredicates) { + ExprWalkerInfo unpushedPreds = mergeChildrenPred(nd, owi, null, false); + return createFilter((Operator)nd, unpushedPreds, owi); + } return null; } @@ -318,20 +352,21 @@ * @param ignoreAliases * @throws SemanticException */ - protected void mergeWithChildrenPred(Node nd, OpWalkerInfo owi, + protected boolean mergeWithChildrenPred(Node nd, OpWalkerInfo owi, ExprWalkerInfo ewi, Set aliases, boolean ignoreAliases) throws SemanticException { + boolean hasUnpushedPredicates = false; if (nd.getChildren() == null || nd.getChildren().size() > 1) { // ppd for multi-insert query is not yet implemented // no-op for leafs - return; + return hasUnpushedPredicates; } Operator op = (Operator) nd; ExprWalkerInfo childPreds = owi .getPrunedPreds((Operator) nd.getChildren() .get(0)); if (childPreds == null) { - return; + return hasUnpushedPredicates; } if (ewi == null) { ewi = new ExprWalkerInfo(); @@ -344,12 +379,41 @@ // input8.q ExprWalkerInfo extractPushdownPreds = ExprWalkerProcFactory .extractPushdownPreds(owi, op, e.getValue()); + if (!extractPushdownPreds.getNonFinalCandidates().isEmpty()) { + hasUnpushedPredicates = true; + } ewi.merge(extractPushdownPreds); logExpr(nd, extractPushdownPreds); } } owi.putPrunedPreds((Operator) nd, ewi); + return hasUnpushedPredicates; } + + protected ExprWalkerInfo mergeChildrenPred(Node nd, OpWalkerInfo owi, + Set excludedAliases, boolean ignoreAliases) + throws SemanticException { + if (nd.getChildren() == null) { + return null; + } + Operator op = (Operator) nd; + ExprWalkerInfo ewi = new ExprWalkerInfo(); + for (Operator child : op.getChildOperators()) { + ExprWalkerInfo childPreds = owi.getPrunedPreds(child); + if (childPreds == null) { + continue; + } + for (Entry> e : childPreds + .getFinalCandidates().entrySet()) { + if (ignoreAliases || excludedAliases == null || + !excludedAliases.contains(e.getKey()) || e.getKey() == null) { + ewi.addPushDowns(e.getKey(), e.getValue()); + logExpr(nd, ewi); + } + } + } + return ewi; + } } protected static Object createFilter(Operator op, @@ -386,7 +450,7 @@ if (condn == null) { return null; } - + if (op instanceof TableScanOperator) { boolean pushFilterToStorage; HiveConf hiveConf = owi.getParseContext().getConf(); @@ -423,6 +487,21 @@ } OpParseContext ctx = new OpParseContext(inputRR); owi.put(output, ctx); + + // remove the candidate filter ops + for (FilterOperator fop : owi.getCandidateFilterOps()) { + List> children = fop.getChildOperators(); + List> parents = fop.getParentOperators(); + for (Operator parent : parents) { + parent.getChildOperators().addAll(children); + parent.removeChild(fop); + } + for (Operator child : children) { + child.getParentOperators().addAll(parents); + child.removeParent(fop); + } + } + owi.getCandidateFilterOps().clear(); return output; } @@ -500,7 +579,7 @@ tableScanDesc.setFilterExpr(decomposed.pushedPredicate); return decomposed.residualPredicate; } - + public static NodeProcessor getFilterProc() { return new FilterPPD(); } @@ -530,7 +609,7 @@ } public static NodeProcessor getUDTFProc() { - return new ScriptPPD(); + return new UDTFPPD(); } public static NodeProcessor getLVFProc() { Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/OpWalkerInfo.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/OpWalkerInfo.java (revision 1038445) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/OpWalkerInfo.java (working copy) @@ -18,9 +18,12 @@ package org.apache.hadoop.hive.ql.ppd; import java.io.Serializable; +import java.util.ArrayList; import java.util.HashMap; +import java.util.List; import java.util.Map; +import org.apache.hadoop.hive.ql.exec.FilterOperator; import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.lib.Node; import org.apache.hadoop.hive.ql.lib.NodeProcessorCtx; @@ -39,11 +42,13 @@ private final HashMap, ExprWalkerInfo> opToPushdownPredMap; private final Map, OpParseContext> opToParseCtxMap; private final ParseContext pGraphContext; + private final List candidateFilterOps; public OpWalkerInfo(ParseContext pGraphContext) { this.pGraphContext = pGraphContext; opToParseCtxMap = pGraphContext.getOpParseCtx(); opToPushdownPredMap = new HashMap, ExprWalkerInfo>(); + candidateFilterOps = new ArrayList(); } public ExprWalkerInfo getPrunedPreds(Operator op) { @@ -67,4 +72,13 @@ public ParseContext getParseContext() { return pGraphContext; } + + public List getCandidateFilterOps() { + return candidateFilterOps; + } + + public void addCandidateFilterOp(FilterOperator fop) { + candidateFilterOps.add(fop); + } + } Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/PredicatePushDown.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/PredicatePushDown.java (revision 1038445) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/PredicatePushDown.java (working copy) @@ -17,13 +17,10 @@ */ package org.apache.hadoop.hive.ql.ppd; -import java.io.Serializable; import java.util.ArrayList; -import java.util.HashMap; import java.util.LinkedHashMap; import java.util.Map; -import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.lib.DefaultGraphWalker; import org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher; import org.apache.hadoop.hive.ql.lib.Dispatcher; @@ -33,7 +30,6 @@ import org.apache.hadoop.hive.ql.lib.Rule; import org.apache.hadoop.hive.ql.lib.RuleRegExp; import org.apache.hadoop.hive.ql.optimizer.Transform; -import org.apache.hadoop.hive.ql.parse.OpParseContext; import org.apache.hadoop.hive.ql.parse.ParseContext; import org.apache.hadoop.hive.ql.parse.SemanticException; @@ -73,12 +69,10 @@ public class PredicatePushDown implements Transform { private ParseContext pGraphContext; - private HashMap, OpParseContext> opToParseCtxMap; @Override public ParseContext transform(ParseContext pctx) throws SemanticException { pGraphContext = pctx; - opToParseCtxMap = pGraphContext.getOpParseCtx(); // create a the context for walking operators OpWalkerInfo opWalkerInfo = new OpWalkerInfo(pGraphContext); Index: ql/src/test/results/clientpositive/auto_join0.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join0.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join0.q.out (working copy) @@ -52,26 +52,22 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [] - 1 [] - Position of Big Table: 0 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [] + 1 [] + Position of Big Table: 0 Stage: Stage-5 Map Reduce @@ -83,46 +79,42 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [] - 1 [] + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [] + 1 [] + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col1, _col2, _col3 - Position of Big Table: 0 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -220,26 +212,22 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [] - 1 [] - Position of Big Table: 1 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [] + 1 [] + Position of Big Table: 1 Stage: Stage-6 Map Reduce @@ -251,46 +239,42 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [] - 1 [] + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [] + 1 [] + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col1, _col2, _col3 - Position of Big Table: 1 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -304,25 +288,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + sort order: + tag: 0 + value expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - sort order: - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string a:src2:src TableScan alias: src @@ -330,25 +310,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + sort order: + tag: 1 + value expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - sort order: - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join11.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join11.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join11.q.out (working copy) @@ -73,41 +73,37 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col0} - 1 {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col0} + 1 {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + outputColumnNames: _col0, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col3 - Position of Big Table: 0 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col3 - Group By Operator - aggregations: - expr: sum(hash(_col0,_col3)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Group By Operator + aggregations: + expr: sum(hash(_col0,_col3)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -159,19 +155,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + Position of Big Table: 1 Stage: Stage-5 Map Reduce @@ -235,22 +227,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src Index: ql/src/test/results/clientpositive/auto_join12.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join12.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join12.q.out (working copy) @@ -82,21 +82,17 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 80) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {_col1} - 2 - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - 2 [Column[_col0]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {_col1} + 2 + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + 2 [Column[_col0]] + Position of Big Table: 0 Stage: Stage-5 Map Reduce @@ -113,44 +109,40 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - Inner Join 0 to 2 - condition expressions: - 0 {_col0} - 1 {_col1} - 2 - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - 2 [Column[_col0]] + Map Join Operator + condition map: + Inner Join 0 to 1 + Inner Join 0 to 2 + condition expressions: + 0 {_col0} + 1 {_col1} + 2 + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + 2 [Column[_col0]] + outputColumnNames: _col0, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col3 - Position of Big Table: 0 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col3 - Group By Operator - aggregations: - expr: sum(hash(_col0,_col3)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Group By Operator + aggregations: + expr: sum(hash(_col0,_col3)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -205,21 +197,17 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {_col1} - 2 - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - 2 [Column[_col0]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {_col1} + 2 + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + 2 [Column[_col0]] + Position of Big Table: 1 src3:src TableScan alias: src @@ -232,21 +220,17 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 80) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {_col1} - 2 - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - 2 [Column[_col0]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {_col1} + 2 + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + 2 [Column[_col0]] + Position of Big Table: 1 Stage: Stage-6 Map Reduce @@ -320,21 +304,17 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {_col1} - 2 - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - 2 [Column[_col0]] - Position of Big Table: 2 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {_col1} + 2 + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + 2 [Column[_col0]] + Position of Big Table: 2 src2:src TableScan alias: src @@ -372,44 +352,40 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 80) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - Inner Join 0 to 2 - condition expressions: - 0 {_col0} - 1 {_col1} - 2 - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - 2 [Column[_col0]] + Map Join Operator + condition map: + Inner Join 0 to 1 + Inner Join 0 to 2 + condition expressions: + 0 {_col0} + 1 {_col1} + 2 + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + 2 [Column[_col0]] + outputColumnNames: _col0, _col3 + Position of Big Table: 2 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col3 - Position of Big Table: 2 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col3 - Group By Operator - aggregations: - expr: sum(hash(_col0,_col3)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Group By Operator + aggregations: + expr: sum(hash(_col0,_col3)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -428,22 +404,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src @@ -478,19 +450,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 80) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 2 Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join13.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join13.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join13.q.out (working copy) @@ -85,28 +85,24 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col0} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - outputColumnNames: _col0, _col2, _col3 - Position of Big Table: 0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col0} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + outputColumnNames: _col0, _col2, _col3 + Position of Big Table: 0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -132,19 +128,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 200) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col3} {_col0} - 1 - handleSkewJoin: false - keys: - 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()] - 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {_col3} {_col0} + 1 + handleSkewJoin: false + keys: + 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()] + 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()] + Position of Big Table: 0 Stage: Stage-6 Map Reduce @@ -246,41 +238,37 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 200) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col3} {_col0} - 1 - handleSkewJoin: false - keys: - 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()] - 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()] - outputColumnNames: _col1, _col2 - Position of Big Table: 1 - Select Operator - expressions: - expr: _col2 - type: string - expr: _col1 - type: string - outputColumnNames: _col2, _col1 - Group By Operator - aggregations: - expr: sum(hash(_col2,_col1)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col3} {_col0} + 1 + handleSkewJoin: false + keys: + 0 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0], Column[_col2]()] + 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[_col0]()] + outputColumnNames: _col1, _col2 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col2 + type: string + expr: _col1 + type: string + outputColumnNames: _col2, _col1 + Group By Operator + aggregations: + expr: sum(hash(_col2,_col1)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -314,19 +302,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 200) - type: boolean - Reduce Output Operator - key expressions: - expr: UDFToDouble(_col0) - type: double - sort order: + - Map-reduce partition columns: - expr: UDFToDouble(_col0) - type: double - tag: 1 + Reduce Output Operator + key expressions: + expr: UDFToDouble(_col0) + type: double + sort order: + + Map-reduce partition columns: + expr: UDFToDouble(_col0) + type: double + tag: 1 Reduce Operator Tree: Join Operator condition map: @@ -375,19 +359,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + Position of Big Table: 1 Stage: Stage-10 Map Reduce @@ -438,22 +418,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src Index: ql/src/test/results/clientpositive/auto_join14.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join14.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join14.q.out (working copy) @@ -38,19 +38,15 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {key} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + Position of Big Table: 0 Stage: Stage-4 Map Reduce @@ -62,44 +58,40 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {key} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - outputColumnNames: _col0, _col5 - Position of Big Table: 0 + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {key} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + outputColumnNames: _col0, _col5 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col5 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -130,33 +122,7 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - Position of Big Table: 1 - - Stage: Stage-5 - Map Reduce - Alias -> Map Operator Tree: - srcpart - TableScan - alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 + HashTable Sink Operator condition expressions: 0 {key} 1 {value} @@ -164,30 +130,48 @@ keys: 0 [Column[key]] 1 [Column[key]] - outputColumnNames: _col0, _col5 Position of Big Table: 1 + + Stage: Stage-5 + Map Reduce + Alias -> Map Operator Tree: + srcpart + TableScan + alias: srcpart + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {key} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + outputColumnNames: _col0, _col5 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col5 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -201,29 +185,6 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - srcpart - TableScan - alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean Reduce Output Operator key expressions: expr: key @@ -232,10 +193,25 @@ Map-reduce partition columns: expr: key type: string - tag: 1 + tag: 0 value expressions: - expr: value + expr: key type: string + srcpart + TableScan + alias: srcpart + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: value + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join16.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join16.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join16.q.out (working copy) @@ -65,56 +65,44 @@ predicate: expr: ((key > 10) and (key > 20)) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 > 20) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col0} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[_col0], Column[_col1]] - 1 [Column[key], Column[value]] - outputColumnNames: _col0, _col3 - Position of Big Table: 0 - Filter Operator - predicate: - expr: (_col3 < 200) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col3 - Group By Operator - aggregations: - expr: sum(hash(_col0,_col3)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col0} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[_col0], Column[_col1]] + 1 [Column[key], Column[value]] + outputColumnNames: _col0, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col3 + Group By Operator + aggregations: + expr: sum(hash(_col0,_col3)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -161,30 +149,22 @@ predicate: expr: ((key > 10) and (key > 20)) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 > 20) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[_col0], Column[_col1]] - 1 [Column[key], Column[value]] - Position of Big Table: 1 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[_col0], Column[_col1]] + 1 [Column[key], Column[value]] + Position of Big Table: 1 Stage: Stage-5 Map Reduce @@ -208,29 +188,25 @@ 1 [Column[key], Column[value]] outputColumnNames: _col0, _col3 Position of Big Table: 1 - Filter Operator - predicate: - expr: (_col3 < 200) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col3 - Group By Operator - aggregations: - expr: sum(hash(_col0,_col3)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col3 + Group By Operator + aggregations: + expr: sum(hash(_col0,_col3)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -244,37 +220,29 @@ predicate: expr: ((key > 10) and (key > 20)) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 > 20) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - expr: _col1 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + sort order: ++ + Map-reduce partition columns: + expr: _col0 + type: string + expr: _col1 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string tab TableScan alias: tab @@ -307,29 +275,25 @@ 1 {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col3 - Filter Operator - predicate: - expr: (_col3 < 200) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col3 - Group By Operator - aggregations: - expr: sum(hash(_col0,_col3)) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col3 + Group By Operator + aggregations: + expr: sum(hash(_col0,_col3)) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/auto_join19.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join19.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join19.q.out (working copy) @@ -68,32 +68,28 @@ 1 [Column[key]] outputColumnNames: _col0, _col2, _col3, _col7 Position of Big Table: 0 - Filter Operator - predicate: - expr: (((_col2 = '2008-04-08') or (_col2 = '2008-04-09')) and ((_col3 = '12') or (_col3 = '11'))) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -148,32 +144,28 @@ 1 [Column[key]] outputColumnNames: _col0, _col2, _col3, _col7 Position of Big Table: 1 - Filter Operator - predicate: - expr: (((_col2 = '2008-04-08') or (_col2 = '2008-04-09')) and ((_col3 = '12') or (_col3 = '11'))) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -223,32 +215,28 @@ 1 {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col2, _col3, _col7 - Filter Operator - predicate: - expr: (((_col2 = '2008-04-08') or (_col2 = '2008-04-09')) and ((_col3 = '12') or (_col3 = '11'))) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 PREHOOK: query: FROM srcpart src1 JOIN src src2 ON (src1.key = src2.key) Index: ql/src/test/results/clientpositive/auto_join20.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join20.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join20.q.out (working copy) @@ -49,25 +49,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 - 1 - 2 {(key < 20)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 + 1 + 2 {(key < 20)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 1 a:src3 TableScan alias: src3 @@ -245,25 +241,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 - 1 - 2 {(key < 20)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 2 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 + 1 + 2 {(key < 20)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 2 a:src2 TableScan alias: src2 @@ -342,24 +334,20 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string a:src2 TableScan alias: src2 @@ -506,25 +494,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 - 1 - 2 {(key < 20)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 + 1 + 2 {(key < 20)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 1 a:src3 TableScan alias: src3 @@ -554,50 +538,46 @@ predicate: expr: (key < 15) type: boolean - Filter Operator - predicate: - expr: (key < 15) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - Right Outer Join0 to 2 - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 - 1 - 2 {(key < 20)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9 - Position of Big Table: 1 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - expr: _col8 - type: string - expr: _col9 - type: string - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Map Join Operator + condition map: + Inner Join 0 to 1 + Right Outer Join0 to 2 + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 + 1 + 2 {(key < 20)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + expr: _col8 + type: string + expr: _col9 + type: string + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -710,25 +690,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 - 1 - 2 {(key < 20)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 2 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 + 1 + 2 {(key < 20)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 2 a:src2 TableScan alias: src2 @@ -736,25 +712,21 @@ predicate: expr: (key < 15) type: boolean - Filter Operator - predicate: - expr: (key < 15) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 - 1 - 2 {(key < 20)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 2 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 + 1 + 2 {(key < 20)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 2 Stage: Stage-7 Map Reduce @@ -815,24 +787,20 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string a:src2 TableScan alias: src2 @@ -840,24 +808,20 @@ predicate: expr: (key < 15) type: boolean - Filter Operator - predicate: - expr: (key < 15) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string a:src3 TableScan alias: src3 Index: ql/src/test/results/clientpositive/auto_join21.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join21.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join21.q.out (working copy) @@ -38,25 +38,21 @@ predicate: expr: (key > 10) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 {(key < 10)} - 1 - 2 {(key < 10)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 {(key < 10)} + 1 + 2 {(key < 10)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 0 src3 TableScan alias: src3 @@ -202,25 +198,21 @@ predicate: expr: (key > 10) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {key} {value} - 2 {key} {value} - filter predicates: - 0 {(key < 10)} - 1 - 2 {(key < 10)} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - Position of Big Table: 2 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {key} {value} + 2 {key} {value} + filter predicates: + 0 {(key < 10)} + 1 + 2 {(key < 10)} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + Position of Big Table: 2 Stage: Stage-6 Map Reduce @@ -298,24 +290,20 @@ predicate: expr: (key > 10) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string src3 TableScan alias: src3 Index: ql/src/test/results/clientpositive/auto_join23.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join23.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join23.q.out (working copy) @@ -67,27 +67,23 @@ 1 [] outputColumnNames: _col0, _col1, _col4, _col5 Position of Big Table: 0 - Filter Operator - predicate: - expr: ((_col0 < 10) and (_col4 < 10)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -171,27 +167,23 @@ 1 [] outputColumnNames: _col0, _col1, _col4, _col5 Position of Big Table: 1 - Filter Operator - predicate: - expr: ((_col0 < 10) and (_col4 < 10)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Local Work: Map Reduce Local Work @@ -237,27 +229,23 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: ((_col0 < 10) and (_col4 < 10)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/auto_join4.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join4.q.out (working copy) @@ -62,26 +62,22 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - Position of Big Table: 0 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + Position of Big Table: 0 Stage: Stage-4 Map Reduce @@ -93,29 +89,36 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Map Join Operator - condition map: - Left Outer Join0 to 1 - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Map Join Operator + condition map: + Left Outer Join0 to 1 + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col1, _col2, _col3 - Position of Big Table: 0 Select Operator expressions: expr: _col0 @@ -129,34 +132,23 @@ outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: - expr: _col0 - type: string + expr: UDFToInteger(_col0) + type: int expr: _col1 type: string - expr: _col2 - type: string + expr: UDFToInteger(_col2) + type: int expr: _col3 type: string outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - expr: UDFToInteger(_col2) - type: int - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -183,31 +175,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -215,31 +203,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join5.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join5.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join5.q.out (working copy) @@ -62,26 +62,22 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - Position of Big Table: 1 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + Position of Big Table: 1 Stage: Stage-4 Map Reduce @@ -93,29 +89,36 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Map Join Operator - condition map: - Right Outer Join0 to 1 - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Map Join Operator + condition map: + Right Outer Join0 to 1 + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string outputColumnNames: _col0, _col1, _col2, _col3 - Position of Big Table: 1 Select Operator expressions: expr: _col0 @@ -129,34 +132,23 @@ outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: - expr: _col0 - type: string + expr: UDFToInteger(_col0) + type: int expr: _col1 type: string - expr: _col2 - type: string + expr: UDFToInteger(_col2) + type: int expr: _col3 type: string outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - expr: UDFToInteger(_col2) - type: int - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -183,31 +175,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -215,31 +203,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join6.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join6.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join6.q.out (working copy) @@ -52,31 +52,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -84,31 +80,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join7.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join7.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join7.q.out (working copy) @@ -62,31 +62,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -94,31 +90,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:c:src3 TableScan alias: src3 @@ -126,31 +118,27 @@ predicate: expr: ((key > 20) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 20) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 2 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/auto_join8.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join8.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join8.q.out (working copy) @@ -62,26 +62,22 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - Position of Big Table: 0 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + Position of Big Table: 0 Stage: Stage-4 Map Reduce @@ -93,29 +89,29 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Map Join Operator - condition map: - Left Outer Join0 to 1 - condition expressions: - 0 {_col0} {_col1} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[_col0]] - outputColumnNames: _col0, _col1, _col2, _col3 - Position of Big Table: 0 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Map Join Operator + condition map: + Left Outer Join0 to 1 + condition expressions: + 0 {_col0} {_col1} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[_col0]] + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Filter Operator + predicate: + expr: (_col2 is null and _col0 is not null) + type: boolean Select Operator expressions: expr: _col0 @@ -127,40 +123,36 @@ expr: _col3 type: string outputColumnNames: _col0, _col1, _col2, _col3 - Filter Operator - predicate: - expr: (_col2 is null and _col0 is not null) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: - expr: _col0 - type: string + expr: UDFToInteger(_col0) + type: int expr: _col1 type: string - expr: _col2 - type: string + expr: UDFToInteger(_col2) + type: int expr: _col3 type: string outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - expr: UDFToInteger(_col2) - type: int - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -187,31 +179,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -219,31 +207,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: @@ -253,21 +237,21 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Filter Operator - predicate: - expr: (_col2 is null and _col0 is not null) - type: boolean + Filter Operator + predicate: + expr: (_col2 is null and _col0 is not null) + type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: expr: _col0 Index: ql/src/test/results/clientpositive/auto_join9.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_join9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/auto_join9.q.out (working copy) @@ -66,32 +66,28 @@ 1 [Column[key]] outputColumnNames: _col0, _col2, _col3, _col7 Position of Big Table: 0 - Filter Operator - predicate: - expr: ((_col2 = '2008-04-08') and (_col3 = '12')) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -146,32 +142,28 @@ 1 [Column[key]] outputColumnNames: _col0, _col2, _col3, _col7 Position of Big Table: 1 - Filter Operator - predicate: - expr: ((_col2 = '2008-04-08') and (_col3 = '12')) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Local Work: Map Reduce Local Work @@ -221,32 +213,28 @@ 1 {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col2, _col3, _col7 - Filter Operator - predicate: - expr: ((_col2 = '2008-04-08') and (_col3 = '12')) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 PREHOOK: query: FROM srcpart src1 JOIN src src2 ON (src1.key = src2.key) Index: ql/src/test/results/clientpositive/bucket2.q.out =================================================================== --- ql/src/test/results/clientpositive/bucket2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucket2.q.out (working copy) @@ -190,28 +190,24 @@ predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string Reduce Operator Tree: Extract File Output Operator Index: ql/src/test/results/clientpositive/bucket3.q.out =================================================================== --- ql/src/test/results/clientpositive/bucket3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucket3.q.out (working copy) @@ -211,36 +211,28 @@ predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Filter Operator - predicate: - expr: (ds = '1') - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - expr: ds - type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 + type: int + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col2 + type: string Reduce Operator Tree: Extract File Output Operator Index: ql/src/test/results/clientpositive/bucket4.q.out =================================================================== --- ql/src/test/results/clientpositive/bucket4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucket4.q.out (working copy) @@ -195,23 +195,19 @@ predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/bucket_groupby.q.out =================================================================== --- ql/src/test/results/clientpositive/bucket_groupby.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucket_groupby.q.out (working copy) @@ -44,36 +44,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '100') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -183,36 +179,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '101') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -291,36 +283,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '101') - type: boolean - Select Operator - expressions: - expr: key - type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: length(key) + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: + expr: length(key) + type: int + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -390,36 +378,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '101') - type: boolean - Select Operator - expressions: - expr: key - type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: abs(length(key)) + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: + expr: abs(length(key)) + type: int + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -491,42 +475,38 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '101') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: key + expr: 3 + type: int + mode: hash + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: 3 + expr: _col1 type: int - mode: hash - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: int - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - expr: _col1 - type: int - tag: -1 - value expressions: - expr: _col2 - type: bigint + sort order: ++ + Map-reduce partition columns: + expr: _col0 + type: string + expr: _col1 + type: int + tag: -1 + value expressions: + expr: _col2 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -607,41 +587,37 @@ subq:clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '101') - type: boolean + Select Operator + expressions: + expr: value + type: string + outputColumnNames: _col0 Select Operator expressions: - expr: value + expr: _col0 type: string outputColumnNames: _col0 - Select Operator - expressions: + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: _col0 type: string - outputColumnNames: _col0 - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -1250,36 +1226,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '102') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -1360,36 +1332,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '102') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: value type: string - outputColumnNames: value - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: value + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -1470,44 +1438,40 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '102') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: key, value + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: expr: key type: string expr: value type: string - outputColumnNames: key, value - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - expr: _col1 - type: string - tag: -1 - value expressions: - expr: _col2 - type: bigint + sort order: ++ + Map-reduce partition columns: + expr: _col0 + type: string + expr: _col1 + type: string + tag: -1 + value expressions: + expr: _col2 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -1637,36 +1601,32 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '103') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: true + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: true - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -1751,44 +1711,40 @@ clustergroupby TableScan alias: clustergroupby - Filter Operator - predicate: - expr: (ds = '103') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + expr: key + type: string + outputColumnNames: value, key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: value type: string expr: key type: string - outputColumnNames: value, key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: value + mode: hash + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: key + expr: _col1 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - expr: _col1 - type: string - tag: -1 - value expressions: - expr: _col2 - type: bigint + sort order: ++ + Map-reduce partition columns: + expr: _col0 + type: string + expr: _col1 + type: string + tag: -1 + value expressions: + expr: _col2 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/bucketmapjoin1.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucketmapjoin1.q.out (working copy) @@ -152,46 +152,41 @@ expr: _col6 type: string outputColumnNames: _col0, _col1, _col5, _col6 - Filter Operator - isSamplingPred: false - predicate: - expr: (_col6 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-32-01_093_6438431130513992258/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-32-01_093_6438431130513992258/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289266321 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_04-43-36_793_2377393369524308875/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_04-43-36_793_2377393369524308875/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289911416 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false @@ -583,50 +578,45 @@ expr: _col6 type: string outputColumnNames: _col0, _col1, _col5, _col6 - Filter Operator - isSamplingPred: false - predicate: - expr: (_col6 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-33-03_655_3362045352453175604/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-33-03_655_3362045352453175604/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - numFiles 1 - numPartitions 0 - numRows 464 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 - transient_lastDdlTime 1289266367 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_04-44-47_227_1478906788720872282/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_04-44-47_227_1478906788720872282/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + numFiles 1 + numPartitions 0 + numRows 464 + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 8983 + transient_lastDdlTime 1289911466 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/bucketmapjoin2.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucketmapjoin2.q.out (working copy) @@ -104,20 +104,15 @@ TableScan alias: b GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + Position of Big Table: 0 Bucket Mapjoin Context: Alias Bucket Base File Name Mapping: b {srcbucket20.txt=[srcbucket22.txt], srcbucket21.txt=[srcbucket23.txt]} @@ -556,23 +551,27 @@ TableScan alias: b GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {key} {value} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {key} {value} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + outputColumnNames: _col0, _col1, _col5 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col5 + type: string outputColumnNames: _col0, _col1, _col5 - Position of Big Table: 1 Select Operator expressions: expr: _col0 @@ -581,46 +580,37 @@ type: string expr: _col5 type: string - outputColumnNames: _col0, _col1, _col5 - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-35-19_461_1296140382333824913/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-35-19_461_1296140382333824913/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - numFiles 1 - numPartitions 0 - numRows 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 - transient_lastDdlTime 1289266501 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_03-24-52_482_6419361805465288120/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_03-24-52_482_6419361805465288120/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + numFiles 1 + numPartitions 0 + numRows 0 + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 + transient_lastDdlTime 1289906667 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/bucketmapjoin3.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucketmapjoin3.q.out (working copy) @@ -104,20 +104,15 @@ TableScan alias: b GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + Position of Big Table: 0 Bucket Mapjoin Context: Alias Bucket Base File Name Mapping: b {srcbucket22.txt=[srcbucket20.txt, srcbucket22.txt], srcbucket23.txt=[srcbucket21.txt, srcbucket23.txt]} @@ -134,23 +129,27 @@ TableScan alias: a GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {key} {value} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {key} {value} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + outputColumnNames: _col0, _col1, _col6 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col6 + type: string outputColumnNames: _col0, _col1, _col6 - Position of Big Table: 0 Select Operator expressions: expr: _col0 @@ -159,42 +158,33 @@ type: string expr: _col6 type: string - outputColumnNames: _col0, _col1, _col6 - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col6 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-36-28_593_5656522145364934245/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-36-28_593_5656522145364934245/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289266588 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_05-20-52_120_3042987158352735578/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_05-20-52_120_3042987158352735578/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289913652 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false @@ -540,20 +530,15 @@ TableScan alias: a GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - Position of Big Table: 1 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + Position of Big Table: 1 Bucket Mapjoin Context: Alias Bucket Base File Name Mapping: a {srcbucket20.txt=[srcbucket22.txt], srcbucket21.txt=[srcbucket23.txt], srcbucket22.txt=[srcbucket22.txt], srcbucket23.txt=[srcbucket23.txt]} @@ -572,23 +557,27 @@ TableScan alias: b GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {key} {value} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {key} {value} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + outputColumnNames: _col0, _col1, _col6 + Position of Big Table: 1 + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col6 + type: string outputColumnNames: _col0, _col1, _col6 - Position of Big Table: 1 Select Operator expressions: expr: _col0 @@ -597,46 +586,37 @@ type: string expr: _col6 type: string - outputColumnNames: _col0, _col1, _col6 - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col6 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-37-28_134_766191094581042721/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_17-37-28_134_766191094581042721/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - numFiles 1 - numPartitions 0 - numRows 564 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 - transient_lastDdlTime 1289266631 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_05-21-57_952_9049475063069442813/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_05-21-57_952_9049475063069442813/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + numFiles 1 + numPartitions 0 + numRows 564 + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 11067 + transient_lastDdlTime 1289913698 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out (working copy) @@ -114,46 +114,41 @@ expr: _col6 type: string outputColumnNames: _col0, _col1, _col5, _col6 - Filter Operator - isSamplingPred: false - predicate: - expr: (_col6 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_18-17-56_639_983307583097560485/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_18-17-56_639_983307583097560485/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289269076 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_05-48-16_859_4190933937441515823/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_05-48-16_859_4190933937441515823/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289915296 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/case_sensitivity.q.out =================================================================== --- ql/src/test/results/clientpositive/case_sensitivity.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/case_sensitivity.q.out (working copy) @@ -33,25 +33,21 @@ predicate: expr: (lint[0] > 0) type: boolean - Filter Operator - predicate: - expr: (lint[0] > 0) - type: boolean - Select Operator - expressions: - expr: lint[1] - type: int - expr: lintstring[0].MYSTRING - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + Select Operator + expressions: + expr: lint[1] + type: int + expr: lintstring[0].MYSTRING + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/cast1.q.out =================================================================== --- ql/src/test/results/clientpositive/cast1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/cast1.q.out (working copy) @@ -31,35 +31,31 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: (3 + 2) - type: int - expr: (3.0 + 2) - type: double - expr: (3 + 2.0) - type: double - expr: (3.0 + 2.0) - type: double - expr: ((3 + UDFToInteger(2.0)) + UDFToInteger(UDFToShort(0))) - type: int - expr: UDFToBoolean(1) - type: boolean - expr: UDFToInteger(true) - type: int - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + Select Operator + expressions: + expr: (3 + 2) + type: int + expr: (3.0 + 2) + type: double + expr: (3 + 2.0) + type: double + expr: (3.0 + 2.0) + type: double + expr: ((3 + UDFToInteger(2.0)) + UDFToInteger(UDFToShort(0))) + type: int + expr: UDFToBoolean(1) + type: boolean + expr: UDFToInteger(true) + type: int + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/cluster.q.out =================================================================== --- ql/src/test/results/clientpositive/cluster.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/cluster.q.out (working copy) @@ -22,31 +22,27 @@ predicate: expr: (key = 10) type: boolean - Filter Operator - predicate: - expr: (key = 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -94,31 +90,27 @@ predicate: expr: (key = 20) type: boolean - Filter Operator - predicate: - expr: (key = 20) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -166,31 +158,27 @@ predicate: expr: (key = 20) type: boolean - Filter Operator - predicate: - expr: (key = 20) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -238,31 +226,27 @@ predicate: expr: (key = 20) type: boolean - Filter Operator - predicate: - expr: (key = 20) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -310,31 +294,27 @@ predicate: expr: (key = 20) type: boolean - Filter Operator - predicate: - expr: (key = 20) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -382,31 +362,27 @@ predicate: expr: (key = 20) type: boolean - Filter Operator - predicate: - expr: (key = 20) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -454,31 +430,27 @@ predicate: expr: (key = 20) type: boolean - Filter Operator - predicate: - expr: (key = 20) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col1 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col1 - type: string - sort order: + - Map-reduce partition columns: - expr: _col1 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -640,25 +612,21 @@ 1 {VALUE._col0} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4 - Filter Operator - predicate: - expr: (_col0 = 20) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce @@ -768,27 +736,23 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: (_col0 = 20) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce @@ -900,27 +864,23 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: (_col0 = 20) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce @@ -1030,25 +990,21 @@ 1 {VALUE._col0} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4 - Filter Operator - predicate: - expr: (_col0 = 20) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce @@ -1129,39 +1085,35 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: expr: _col0 type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string null-subquery2:unioninput-subquery2:src TableScan alias: src @@ -1169,39 +1121,35 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: expr: _col0 type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string Reduce Operator Tree: Extract File Output Operator Index: ql/src/test/results/clientpositive/combine2.q.out =================================================================== --- ql/src/test/results/clientpositive/combine2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/combine2.q.out (working copy) @@ -86,28 +86,24 @@ combine2 TableScan alias: combine2 - Filter Operator - predicate: - expr: value is not null - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + tag: -1 + value expressions: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -192,24 +188,19 @@ TableScan alias: combine2 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: value is not null - type: boolean - Select Operator - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: bigint + Select Operator + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: bigint Needs Tagging: false Path -> Alias: pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/combine2/value=2010-04-21 09%3A45%3A00 [combine2] @@ -697,36 +688,32 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: ds is not null - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: ds + type: string + outputColumnNames: ds + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: ds type: string - outputColumnNames: ds - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: ds + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/create_view.q.out =================================================================== --- ql/src/test/results/clientpositive/create_view.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/create_view.q.out (working copy) @@ -194,23 +194,19 @@ expr: value type: string outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 = 18) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/disable_merge_for_bucketing.q.out =================================================================== --- ql/src/test/results/clientpositive/disable_merge_for_bucketing.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/disable_merge_for_bucketing.q.out (working copy) @@ -190,28 +190,24 @@ predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string Reduce Operator Tree: Extract File Output Operator Index: ql/src/test/results/clientpositive/filter_join_breaktask.q.out =================================================================== --- ql/src/test/results/clientpositive/filter_join_breaktask.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/filter_join_breaktask.q.out (working copy) @@ -48,28 +48,18 @@ predicate: expr: key is not null type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: key is not null - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: int - sort order: + - Map-reduce partition columns: - expr: key - type: int - tag: 0 - value expressions: - expr: key - type: int + Reduce Output Operator + key expressions: + expr: key + type: int + sort order: + + Map-reduce partition columns: + expr: key + type: int + tag: 0 + value expressions: + expr: key + type: int m TableScan alias: m @@ -79,25 +69,20 @@ predicate: expr: (value is not null and (value <> '')) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: int - sort order: + - Map-reduce partition columns: - expr: key - type: int - tag: 1 - value expressions: - expr: value - type: string - expr: ds - type: string + Reduce Output Operator + key expressions: + expr: key + type: int + sort order: + + Map-reduce partition columns: + expr: key + type: int + tag: 1 + value expressions: + expr: value + type: string + expr: ds + type: string Needs Tagging: true Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08 [f, m] @@ -159,36 +144,21 @@ 1 {VALUE._col1} {VALUE._col2} handleSkewJoin: false outputColumnNames: _col0, _col6, _col7 - Filter Operator - isSamplingPred: false - predicate: - expr: (_col7 = '2008-04-08') - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: _col6 is not null - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (_col6 <> '') - type: boolean - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/sdong/hive_2010-11-03_11-50-07_978_4040810463207546064/-mr-10002 - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - columns _col0,_col6 - columns.types int,string - escape.delim \ - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-16_03-54-15_497_5452320809327694970/-mr-10002 + NumFilesPerFileSink: 1 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + columns _col0,_col6,_col7 + columns.types int,string,string + escape.delim \ + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-2 Map Reduce @@ -210,23 +180,18 @@ TableScan alias: g GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: value - type: string - sort order: + - Map-reduce partition columns: - expr: value - type: string - tag: 1 - value expressions: - expr: value - type: string + Reduce Output Operator + key expressions: + expr: value + type: string + sort order: + + Map-reduce partition columns: + expr: value + type: string + tag: 1 + value expressions: + expr: value + type: string Needs Tagging: true Path -> Alias: file:/tmp/sdong/hive_2010-11-03_11-50-07_978_4040810463207546064/-mr-10002 [$INTNAME] @@ -238,15 +203,15 @@ input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: - columns _col0,_col6 - columns.types int,string + columns _col0,_col6,_col7 + columns.types int,string,string escape.delim \ input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: - columns _col0,_col6 - columns.types int,string + columns _col0,_col6,_col7 + columns.types int,string,string escape.delim \ pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/filter_join_breaktask/ds=2008-04-08 Partition Index: ql/src/test/results/clientpositive/groupby_map_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/groupby_map_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/groupby_map_ppr.q.out (working copy) @@ -33,46 +33,41 @@ TableScan alias: src GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: key, value + Group By Operator + aggregations: + expr: count(DISTINCT substr(value, 5)) + expr: sum(substr(value, 5)) + bucketGroup: false + keys: + expr: substr(key, 1, 1) type: string - expr: value + expr: substr(value, 5) type: string - outputColumnNames: key, value - Group By Operator - aggregations: - expr: count(DISTINCT substr(value, 5)) - expr: sum(substr(value, 5)) - bucketGroup: false - keys: - expr: substr(key, 1, 1) + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: substr(value, 5) + expr: _col1 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col2 - type: bigint - expr: _col3 - type: double + sort order: ++ + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col2 + type: bigint + expr: _col3 + type: double Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src] Index: ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out =================================================================== --- ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out (working copy) @@ -33,56 +33,51 @@ TableScan alias: src GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: key, value + Group By Operator + aggregations: + expr: count(DISTINCT substr(value, 5)) + expr: sum(substr(value, 5)) + expr: sum(DISTINCT substr(value, 5)) + expr: count(DISTINCT value) + bucketGroup: false + keys: + expr: substr(key, 1, 1) type: string + expr: substr(value, 5) + type: string expr: value type: string - outputColumnNames: key, value - Group By Operator - aggregations: - expr: count(DISTINCT substr(value, 5)) - expr: sum(substr(value, 5)) - expr: sum(DISTINCT substr(value, 5)) - expr: count(DISTINCT value) - bucketGroup: false - keys: - expr: substr(key, 1, 1) + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: substr(value, 5) + expr: _col1 type: string - expr: value + expr: _col2 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - sort order: +++ - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col3 - type: bigint - expr: _col4 - type: double - expr: _col5 - type: double - expr: _col6 - type: bigint + sort order: +++ + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col3 + type: bigint + expr: _col4 + type: double + expr: _col5 + type: double + expr: _col6 + type: bigint Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src] Index: ql/src/test/results/clientpositive/groupby_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/groupby_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/groupby_ppr.q.out (working copy) @@ -33,29 +33,24 @@ TableScan alias: src GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: key, value + Reduce Output Operator + key expressions: + expr: substr(key, 1, 1) type: string - expr: value + expr: substr(value, 5) type: string - outputColumnNames: key, value - Reduce Output Operator - key expressions: - expr: substr(key, 1, 1) - type: string - expr: substr(value, 5) - type: string - sort order: ++ - Map-reduce partition columns: - expr: substr(key, 1, 1) - type: string - tag: -1 + sort order: ++ + Map-reduce partition columns: + expr: substr(key, 1, 1) + type: string + tag: -1 Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src] Index: ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out =================================================================== --- ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out (working copy) @@ -33,31 +33,26 @@ TableScan alias: src GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: key, value + Reduce Output Operator + key expressions: + expr: substr(key, 1, 1) type: string + expr: substr(value, 5) + type: string expr: value type: string - outputColumnNames: key, value - Reduce Output Operator - key expressions: - expr: substr(key, 1, 1) - type: string - expr: substr(value, 5) - type: string - expr: value - type: string - sort order: +++ - Map-reduce partition columns: - expr: substr(key, 1, 1) - type: string - tag: -1 + sort order: +++ + Map-reduce partition columns: + expr: substr(key, 1, 1) + type: string + tag: -1 Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [src] Index: ql/src/test/results/clientpositive/implicit_cast1.q.out =================================================================== --- ql/src/test/results/clientpositive/implicit_cast1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/implicit_cast1.q.out (working copy) @@ -31,23 +31,19 @@ predicate: expr: (a <> 0) type: boolean - Filter Operator - predicate: - expr: (a <> 0) - type: boolean - Select Operator - expressions: - expr: a - type: bigint - expr: b - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: a + type: bigint + expr: b + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/input11.q.out =================================================================== --- ql/src/test/results/clientpositive/input11.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input11.q.out (working copy) @@ -33,32 +33,28 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: key + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: value - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/input11_limit.q.out =================================================================== --- ql/src/test/results/clientpositive/input11_limit.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input11_limit.q.out (working copy) @@ -30,26 +30,22 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Limit - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Limit + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract Limit Index: ql/src/test/results/clientpositive/input14.q.out =================================================================== --- ql/src/test/results/clientpositive/input14.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input14.q.out (working copy) @@ -48,48 +48,48 @@ output info: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Filter Operator + predicate: + expr: (_col0 < 100) + type: boolean + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 - type: string + expr: UDFToInteger(_col0) + type: int expr: _col1 type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/input18.q.out =================================================================== --- ql/src/test/results/clientpositive/input18.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input18.q.out (working copy) @@ -52,48 +52,48 @@ output info: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Filter Operator + predicate: + expr: (_col0 < 100) + type: boolean + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: regexp_replace(_col1, ' ', '+') + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: regexp_replace(_col1, ' ', '+') - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/input23.q.out =================================================================== --- ql/src/test/results/clientpositive/input23.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input23.q.out (working copy) @@ -106,47 +106,42 @@ 1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} handleSkewJoin: false outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7, _col8, _col9 - Filter Operator - isSamplingPred: false - predicate: - expr: ((((_col2 = '2008-04-08') and (_col3 = '11')) and (_col8 = '2008-04-08')) and (_col9 = '14')) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - expr: _col6 - type: string - expr: _col7 - type: string - expr: _col8 - type: string - expr: _col9 - type: string - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/sdong/hive_2010-10-29_15-29-06_211_4226521001698904551/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/sdong/hive_2010-10-29_15-29-06_211_4226521001698904551/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 - columns.types string:string:string:string:string:string:string:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + expr: _col6 + type: string + expr: _col7 + type: string + expr: _col8 + type: string + expr: _col9 + type: string + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-16_23-35-46_527_104890476863280364/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-16_23-35-46_527_104890476863280364/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 + columns.types string:string:string:string:string:string:string:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/input24.q.out =================================================================== --- ql/src/test/results/clientpositive/input24.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input24.q.out (working copy) @@ -30,23 +30,19 @@ x TableScan alias: x - Filter Operator - predicate: - expr: (d = '2009-01-01') - type: boolean - Select Operator - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: bigint + Select Operator + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/input25.q.out =================================================================== --- ql/src/test/results/clientpositive/input25.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input25.q.out (working copy) @@ -47,30 +47,26 @@ null-subquery1:subq-subquery1:x TableScan alias: x - Filter Operator - predicate: - expr: (d = '2009-01-01') - type: boolean - Select Operator - expressions: - expr: a - type: int - expr: b - type: int - expr: d - type: string - outputColumnNames: _col0, _col1, _col2 - Limit - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: int - expr: _col2 - type: string + Select Operator + expressions: + expr: a + type: int + expr: b + type: int + expr: d + type: string + outputColumnNames: _col0, _col1, _col2 + Limit + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: int + expr: _col2 + type: string Reduce Operator Tree: Extract Limit @@ -125,30 +121,26 @@ null-subquery2:subq-subquery2:x TableScan alias: x - Filter Operator - predicate: - expr: (d = '2009-02-02') - type: boolean - Select Operator - expressions: - expr: a - type: int - expr: b - type: int - expr: d - type: string - outputColumnNames: _col0, _col1, _col2 - Limit - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: int - expr: _col2 - type: string + Select Operator + expressions: + expr: a + type: int + expr: b + type: int + expr: d + type: string + outputColumnNames: _col0, _col1, _col2 + Limit + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: int + expr: _col2 + type: string Reduce Operator Tree: Extract Limit Index: ql/src/test/results/clientpositive/input26.q.out =================================================================== --- ql/src/test/results/clientpositive/input26.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input26.q.out (working copy) @@ -28,36 +28,32 @@ null-subquery1:subq-subquery1:a TableScan alias: a - Filter Operator - predicate: - expr: ((ds = '2008-04-08') and (hr = '11')) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + tag: -1 + value expressions: + expr: _col0 type: string - expr: ds + expr: _col1 type: string - expr: hr + expr: _col2 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + expr: _col3 + type: string Reduce Operator Tree: Extract Limit @@ -120,34 +116,30 @@ predicate: expr: ((ds = '2008-04-08') and (hr = '14')) type: boolean - Filter Operator - predicate: - expr: ((ds = '2008-04-08') and (hr = '14')) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Limit - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Limit + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string Reduce Operator Tree: Extract Limit Index: ql/src/test/results/clientpositive/input2_limit.q.out =================================================================== --- ql/src/test/results/clientpositive/input2_limit.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input2_limit.q.out (working copy) @@ -22,24 +22,20 @@ predicate: expr: (key < 300) type: boolean - Filter Operator - predicate: - expr: (key < 300) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/input31.q.out =================================================================== --- ql/src/test/results/clientpositive/input31.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input31.q.out (working copy) @@ -35,23 +35,19 @@ predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: bigint + Select Operator + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/input39.q.out =================================================================== --- ql/src/test/results/clientpositive/input39.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input39.q.out (working copy) @@ -121,23 +121,19 @@ 1 {VALUE._col2} handleSkewJoin: false outputColumnNames: _col2, _col7 - Filter Operator - predicate: - expr: ((_col2 = '1') and (_col7 = '1')) - type: boolean - Select Operator - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce Index: ql/src/test/results/clientpositive/input42.q.out =================================================================== --- ql/src/test/results/clientpositive/input42.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input42.q.out (working copy) @@ -19,39 +19,34 @@ TableScan alias: a GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col3 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 type: string - expr: hr + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col3 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + expr: _col2 + type: string + expr: _col3 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [a] @@ -1199,39 +1194,34 @@ predicate: expr: (key < 200) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '2008-04-08') and (key < 200)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col3 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 type: string - expr: hr + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col3 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + expr: _col2 + type: string + expr: _col3 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [a] Index: ql/src/test/results/clientpositive/input6.q.out =================================================================== --- ql/src/test/results/clientpositive/input6.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input6.q.out (working copy) @@ -33,25 +33,21 @@ predicate: expr: key is null type: boolean - Filter Operator - predicate: - expr: key is null - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/input9.q.out =================================================================== --- ql/src/test/results/clientpositive/input9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input9.q.out (working copy) @@ -33,32 +33,28 @@ predicate: expr: (null = null) type: boolean - Filter Operator - predicate: - expr: (null = null) - type: boolean + Select Operator + expressions: + expr: null + type: string + expr: key + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: null - type: string - expr: key - type: string + expr: _col0 + type: void + expr: UDFToInteger(_col1) + type: int outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: _col0 - type: void - expr: UDFToInteger(_col1) - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/input_part1.q.out =================================================================== --- ql/src/test/results/clientpositive/input_part1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_part1.q.out (working copy) @@ -35,59 +35,54 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((key < 100) and (ds = '2008-04-08')) and (hr = '12')) - type: boolean + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: hr + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: - expr: key + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: value + expr: _col2 type: string - expr: hr + expr: _col3 type: string - expr: ds - type: string outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/sdong/www/hive-trunk/build/ql/scratchdir/hive_2010-10-29_15-33-38_542_4254366096139103837/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/sdong/www/hive-trunk/build/ql/scratchdir/hive_2010-10-29_15-33-38_542_4254366096139103837/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,hr,ds - columns.types int:string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1288391618 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-43-28_125_8691432559784715360/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-43-28_125_8691432559784715360/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,hr,ds + columns.types int:string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289979808 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [srcpart] Index: ql/src/test/results/clientpositive/input_part5.q.out =================================================================== --- ql/src/test/results/clientpositive/input_part5.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_part5.q.out (working copy) @@ -33,29 +33,25 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: ((ds = '2008-04-08') and (key < 100)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: tmptable + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: tmptable Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/input_part6.q.out =================================================================== --- ql/src/test/results/clientpositive/input_part6.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_part6.q.out (working copy) @@ -22,28 +22,24 @@ predicate: expr: (ds = ((2008 - 4) - 8)) type: boolean - Filter Operator - predicate: - expr: (ds = ((2008 - 4) - 8)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/input_part7.q.out =================================================================== --- ql/src/test/results/clientpositive/input_part7.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_part7.q.out (working copy) @@ -34,49 +34,44 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '2008-04-08') and (key < 100)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Union - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: expr: _col0 type: string + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: string expr: _col1 type: string expr: _col2 type: string expr: _col3 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string null-subquery2:a-subquery2:y TableScan alias: y @@ -86,49 +81,44 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '2008-04-08') and (key < 100)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Union - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: expr: _col0 type: string + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: string expr: _col1 type: string expr: _col2 type: string expr: _col3 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [null-subquery1:a-subquery1:x, null-subquery2:a-subquery2:y] Index: ql/src/test/results/clientpositive/input_part9.q.out =================================================================== --- ql/src/test/results/clientpositive/input_part9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_part9.q.out (working copy) @@ -24,39 +24,34 @@ predicate: expr: key is not null type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key is not null and (ds = '2008-04-08')) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col3 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 type: string - expr: hr + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col3 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + expr: _col2 + type: string + expr: _col3 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [x] Index: ql/src/test/results/clientpositive/input_testxpath2.q.out =================================================================== --- ql/src/test/results/clientpositive/input_testxpath2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_testxpath2.q.out (working copy) @@ -33,27 +33,23 @@ predicate: expr: (lint is not null and (not mstringstring is null)) type: boolean - Filter Operator - predicate: - expr: (lint is not null and (not mstringstring is null)) - type: boolean - Select Operator - expressions: - expr: size(lint) - type: int - expr: size(lintstring) - type: int - expr: size(mstringstring) - type: int - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + Select Operator + expressions: + expr: size(lint) + type: int + expr: size(lintstring) + type: int + expr: size(mstringstring) + type: int + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/input_testxpath4.q.out =================================================================== --- ql/src/test/results/clientpositive/input_testxpath4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/input_testxpath4.q.out (working copy) @@ -107,23 +107,19 @@ predicate: expr: ((mstringstring['key_9'] is not null and lintstring.myint is not null) and lintstring is not null) type: boolean - Filter Operator - predicate: - expr: ((mstringstring['key_9'] is not null and lintstring.myint is not null) and lintstring is not null) - type: boolean - Select Operator - expressions: - expr: mstringstring['key_9'] - type: string - expr: lintstring.myint - type: array - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: mstringstring['key_9'] + type: string + expr: lintstring.myint + type: array + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/join0.q.out =================================================================== --- ql/src/test/results/clientpositive/join0.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join0.q.out (working copy) @@ -33,25 +33,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + sort order: + tag: 0 + value expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - sort order: - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string src2:src TableScan alias: src @@ -59,25 +55,21 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + sort order: + tag: 1 + value expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - sort order: - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join11.q.out =================================================================== --- ql/src/test/results/clientpositive/join11.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join11.q.out (working copy) @@ -37,22 +37,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src Index: ql/src/test/results/clientpositive/join12.q.out =================================================================== --- ql/src/test/results/clientpositive/join12.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join12.q.out (working copy) @@ -43,22 +43,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src @@ -93,19 +89,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 80) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 2 Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join13.q.out =================================================================== --- ql/src/test/results/clientpositive/join13.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join13.q.out (working copy) @@ -44,22 +44,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src @@ -130,19 +126,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 200) - type: boolean - Reduce Output Operator - key expressions: - expr: UDFToDouble(_col0) - type: double - sort order: + - Map-reduce partition columns: - expr: UDFToDouble(_col0) - type: double - tag: 1 + Reduce Output Operator + key expressions: + expr: UDFToDouble(_col0) + type: double + sort order: + + Map-reduce partition columns: + expr: UDFToDouble(_col0) + type: double + tag: 1 Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join14.q.out =================================================================== --- ql/src/test/results/clientpositive/join14.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join14.q.out (working copy) @@ -30,29 +30,6 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - srcpart - TableScan - alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean Reduce Output Operator key expressions: expr: key @@ -61,10 +38,25 @@ Map-reduce partition columns: expr: key type: string - tag: 1 + tag: 0 value expressions: - expr: value + expr: key type: string + srcpart + TableScan + alias: srcpart + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: value + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join16.q.out =================================================================== --- ql/src/test/results/clientpositive/join16.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join16.q.out (working copy) @@ -20,37 +20,29 @@ predicate: expr: ((key > 10) and (key > 20)) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 > 20) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - expr: _col1 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + sort order: ++ + Map-reduce partition columns: + expr: _col0 + type: string + expr: _col1 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string tab TableScan alias: tab @@ -83,23 +75,19 @@ 1 {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col3 - Filter Operator - predicate: - expr: (_col3 < 200) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/join19.q.out =================================================================== --- ql/src/test/results/clientpositive/join19.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join19.q.out (working copy) @@ -135,27 +135,23 @@ predicate: expr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Citation')) type: boolean - Filter Operator - predicate: - expr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Citation')) - type: boolean - Select Operator - expressions: - expr: subject + Select Operator + expressions: + expr: subject + type: string + outputColumnNames: _col0 + Reduce Output Operator + key expressions: + expr: _col0 type: string - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string t22:t2 TableScan alias: t2 @@ -163,29 +159,25 @@ predicate: expr: (predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') type: boolean - Filter Operator - predicate: - expr: (predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') - type: boolean - Select Operator - expressions: - expr: subject + Select Operator + expressions: + expr: subject + type: string + expr: object + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: object + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col1 + type: string t33:t3 TableScan alias: t3 @@ -193,29 +185,25 @@ predicate: expr: (predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_from') type: boolean - Filter Operator - predicate: - expr: (predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_from') - type: boolean - Select Operator - expressions: - expr: subject + Select Operator + expressions: + expr: subject + type: string + expr: object + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col1 type: string - expr: object + sort order: + + Map-reduce partition columns: + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col1 - type: string - sort order: + - Map-reduce partition columns: - expr: _col1 - type: string - tag: 2 - value expressions: - expr: _col0 - type: string + tag: 2 + value expressions: + expr: _col0 + type: string Reduce Operator Tree: Join Operator condition map: @@ -261,24 +249,20 @@ predicate: expr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Author')) type: boolean - Filter Operator - predicate: - expr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Author')) - type: boolean - Select Operator - expressions: - expr: subject + Select Operator + expressions: + expr: subject + type: string + outputColumnNames: _col0 + Reduce Output Operator + key expressions: + expr: _col0 type: string - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 1 t55:t5 TableScan alias: t5 @@ -286,29 +270,25 @@ predicate: expr: (predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_to') type: boolean - Filter Operator - predicate: - expr: (predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_to') - type: boolean - Select Operator - expressions: - expr: subject + Select Operator + expressions: + expr: subject + type: string + expr: object + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: object + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 - value expressions: - expr: _col1 - type: string + tag: 2 + value expressions: + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: @@ -356,29 +336,25 @@ predicate: expr: (predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') type: boolean - Filter Operator - predicate: - expr: (predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') - type: boolean - Select Operator - expressions: - expr: subject + Select Operator + expressions: + expr: subject + type: string + expr: object + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: object + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join20.q.out =================================================================== --- ql/src/test/results/clientpositive/join20.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join20.q.out (working copy) @@ -25,24 +25,20 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string src2 TableScan alias: src2 @@ -746,24 +742,20 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string src2 TableScan alias: src2 @@ -771,24 +763,20 @@ predicate: expr: (key < 15) type: boolean - Filter Operator - predicate: - expr: (key < 15) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string src3 TableScan alias: src3 Index: ql/src/test/results/clientpositive/join21.q.out =================================================================== --- ql/src/test/results/clientpositive/join21.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join21.q.out (working copy) @@ -40,24 +40,20 @@ predicate: expr: (key > 10) type: boolean - Filter Operator - predicate: - expr: (key > 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string src3 TableScan alias: src3 Index: ql/src/test/results/clientpositive/join23.q.out =================================================================== --- ql/src/test/results/clientpositive/join23.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join23.q.out (working copy) @@ -55,27 +55,23 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: ((_col0 < 10) and (_col4 < 10)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce Index: ql/src/test/results/clientpositive/join26.q.out =================================================================== --- ql/src/test/results/clientpositive/join26.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join26.q.out (working copy) @@ -76,75 +76,65 @@ TableScan alias: z GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (hr = 11) - type: boolean - Map Join Operator - condition map: - Inner Join 0 to 1 - Inner Join 0 to 2 - condition expressions: - 0 {key} - 1 {value} - 2 {value} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[key]] - 2 [Column[key]] - outputColumnNames: _col0, _col5, _col9 - Position of Big Table: 2 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col5 - type: string - expr: _col9 - type: string - outputColumnNames: _col0, _col5, _col9 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col9 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_18-59-23_623_4518960640790293704/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_18-59-23_623_4518960640790293704/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,val2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 - name dest_j1 - serialization.ddl struct dest_j1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289271563 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest_j1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Map Join Operator + condition map: + Inner Join 0 to 1 + Inner Join 0 to 2 + condition expressions: + 0 {key} + 1 {value} + 2 {value} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[key]] + 2 [Column[key]] + outputColumnNames: _col0, _col5, _col9 + Position of Big Table: 2 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col5 + type: string + expr: _col9 + type: string + outputColumnNames: _col0, _col5, _col9 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col9 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-51-11_306_6062456690049056800/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-51-11_306_6062456690049056800/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,val2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 + name dest_j1 + serialization.ddl struct dest_j1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289980271 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest_j1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/join28.q.out =================================================================== --- ql/src/test/results/clientpositive/join28.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join28.q.out (working copy) @@ -57,23 +57,15 @@ z TableScan alias: z - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - predicate: - expr: (hr = 11) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[key]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce Index: ql/src/test/results/clientpositive/join32.q.out =================================================================== --- ql/src/test/results/clientpositive/join32.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join32.q.out (working copy) @@ -139,25 +139,15 @@ TableScan alias: z GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (hr = 11) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col5} {_col0} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[_col1]] - 1 [Column[value]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {_col5} {_col0} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[_col1]] + 1 [Column[value]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce Index: ql/src/test/results/clientpositive/join33.q.out =================================================================== --- ql/src/test/results/clientpositive/join33.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join33.q.out (working copy) @@ -155,28 +155,18 @@ TableScan alias: z GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (hr = 11) - type: boolean - Reduce Output Operator - key expressions: - expr: value - type: string - sort order: + - Map-reduce partition columns: - expr: value - type: string - tag: 1 - value expressions: - expr: value - type: string + Reduce Output Operator + key expressions: + expr: value + type: string + sort order: + + Map-reduce partition columns: + expr: value + type: string + tag: 1 + value expressions: + expr: value + type: string Needs Tagging: true Path -> Alias: file:/tmp/liyintang/hive_2010-11-08_19-01-43_620_231968462576941228/-mr-10002 [file:/tmp/liyintang/hive_2010-11-08_19-01-43_620_231968462576941228/-mr-10002] Index: ql/src/test/results/clientpositive/join34.q.out =================================================================== --- ql/src/test/results/clientpositive/join34.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join34.q.out (working copy) @@ -69,75 +69,70 @@ predicate: expr: (key < 20) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key < 20) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col1} - 1 {key} {value} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[key]] + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col1} + 1 {key} {value} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[key]] + outputColumnNames: _col1, _col2, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 Select Operator expressions: - expr: _col1 - type: string expr: _col2 type: string expr: _col3 type: string - outputColumnNames: _col1, _col2, _col3 - Select Operator - expressions: - expr: _col2 - type: string - expr: _col3 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-04_12-12-19_150_396302363179050554/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-04_12-12-19_150_396302363179050554/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,val2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 - name dest_j1 - serialization.ddl struct dest_j1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1288897939 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest_j1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + expr: _col1 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-54-53_787_7989208913285946680/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-54-53_787_7989208913285946680/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,val2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 + name dest_j1 + serialization.ddl struct dest_j1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289980493 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest_j1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false null-subquery2:subq1-subquery2:x1 TableScan alias: x1 @@ -147,75 +142,70 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Map Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {_col1} - 1 {key} {value} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[key]] + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Map Join Operator + condition map: + Inner Join 0 to 1 + condition expressions: + 0 {_col1} + 1 {key} {value} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[key]] + outputColumnNames: _col1, _col2, _col3 + Position of Big Table: 0 + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 Select Operator expressions: - expr: _col1 - type: string expr: _col2 type: string expr: _col3 type: string - outputColumnNames: _col1, _col2, _col3 - Select Operator - expressions: - expr: _col2 - type: string - expr: _col3 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-04_12-12-19_150_396302363179050554/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-04_12-12-19_150_396302363179050554/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,val2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 - name dest_j1 - serialization.ddl struct dest_j1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1288897939 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest_j1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + expr: _col1 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-54-53_787_7989208913285946680/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-16_23-54-53_787_7989208913285946680/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,val2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 + name dest_j1 + serialization.ddl struct dest_j1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289980493 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest_j1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/join35.q.out =================================================================== --- ql/src/test/results/clientpositive/join35.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join35.q.out (working copy) @@ -50,37 +50,32 @@ predicate: expr: (key < 20) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key < 20) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Needs Tagging: false Path -> Alias: pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/src [null-subquery1:subq1-subquery1:x] @@ -472,37 +467,32 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Needs Tagging: false Path -> Alias: pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/src [null-subquery2:subq1-subquery2:x1] Index: ql/src/test/results/clientpositive/join38.q.out =================================================================== --- ql/src/test/results/clientpositive/join38.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join38.q.out (working copy) @@ -143,44 +143,40 @@ expr: _col15 type: string outputColumnNames: _col1, _col9, _col15 - Filter Operator - predicate: - expr: (_col15 = 111) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: _col1 + type: string + expr: _col9 + type: string + outputColumnNames: _col1, _col9 + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: _col1 type: string expr: _col9 type: string - outputColumnNames: _col1, _col9 - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: + mode: hash + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 + type: string expr: _col1 type: string - expr: _col9 + sort order: ++ + Map-reduce partition columns: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - Map-reduce partition columns: - expr: _col0 - type: string - expr: _col1 - type: string - tag: -1 - value expressions: - expr: _col2 - type: bigint + expr: _col1 + type: string + tag: -1 + value expressions: + expr: _col2 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/join39.q.out =================================================================== --- ql/src/test/results/clientpositive/join39.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join39.q.out (working copy) @@ -40,26 +40,22 @@ predicate: expr: (key <= 100) type: boolean - Filter Operator - predicate: - expr: (key <= 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[_col0]] - Position of Big Table: 0 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[_col0]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce Index: ql/src/test/results/clientpositive/join4.q.out =================================================================== --- ql/src/test/results/clientpositive/join4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join4.q.out (working copy) @@ -52,31 +52,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -84,31 +80,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join40.q.out =================================================================== --- ql/src/test/results/clientpositive/join40.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join40.q.out (working copy) @@ -39,31 +39,27 @@ predicate: expr: (key <= 100) type: boolean - Filter Operator - predicate: - expr: (key <= 100) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: @@ -1812,24 +1808,20 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string src2 TableScan alias: src2 @@ -2533,24 +2525,20 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string src2 TableScan alias: src2 @@ -2558,24 +2546,20 @@ predicate: expr: (key < 15) type: boolean - Filter Operator - predicate: - expr: (key < 15) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string src3 TableScan alias: src3 @@ -3266,26 +3250,22 @@ predicate: expr: (key <= 100) type: boolean - Filter Operator - predicate: - expr: (key <= 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - HashTable Sink Operator - condition expressions: - 0 {key} {value} - 1 {_col0} {_col1} - handleSkewJoin: false - keys: - 0 [Column[key]] - 1 [Column[_col0]] - Position of Big Table: 0 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + HashTable Sink Operator + condition expressions: + 0 {key} {value} + 1 {_col0} {_col1} + handleSkewJoin: false + keys: + 0 [Column[key]] + 1 [Column[_col0]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce Index: ql/src/test/results/clientpositive/join5.q.out =================================================================== --- ql/src/test/results/clientpositive/join5.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join5.q.out (working copy) @@ -52,31 +52,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -84,31 +80,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join6.q.out =================================================================== --- ql/src/test/results/clientpositive/join6.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join6.q.out (working copy) @@ -52,31 +52,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -84,31 +80,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join7.q.out =================================================================== --- ql/src/test/results/clientpositive/join7.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join7.q.out (working copy) @@ -62,31 +62,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -94,31 +90,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:c:src3 TableScan alias: src3 @@ -126,31 +118,27 @@ predicate: expr: ((key > 20) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 20) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 2 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/join8.q.out =================================================================== --- ql/src/test/results/clientpositive/join8.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join8.q.out (working copy) @@ -52,31 +52,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -84,31 +80,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: @@ -118,21 +110,21 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Filter Operator - predicate: - expr: (_col2 is null and _col0 is not null) - type: boolean + Filter Operator + predicate: + expr: (_col2 is null and _col0 is not null) + type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: expr: _col0 Index: ql/src/test/results/clientpositive/join9.q.out =================================================================== --- ql/src/test/results/clientpositive/join9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join9.q.out (working copy) @@ -150,51 +150,46 @@ 1 {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col2, _col3, _col7 - Filter Operator - isSamplingPred: false - predicate: - expr: ((_col2 = '2008-04-08') and (_col3 = '12')) - type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col7 + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: _col0 + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: _col7 - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/sdong/www/hive-trunk/build/ql/scratchdir/hive_2010-10-29_15-41-58_075_5780284635323173562/-ext-10000 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/sdong/www/hive-trunk/build/ql/scratchdir/hive_2010-10-29_15-41-58_075_5780284635323173562/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1288392118 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_00-00-25_037_6563073372084295230/-ext-10000 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_00-00-25_037_6563073372084295230/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289980825 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/join_map_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/join_map_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/join_map_ppr.q.out (working copy) @@ -106,53 +106,48 @@ expr: _col11 type: string outputColumnNames: _col0, _col5, _col9, _col10, _col11 - Filter Operator - isSamplingPred: false - predicate: - expr: ((_col10 = '2008-04-08') and (_col11 = 11)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col9 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,val2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 - name dest_j1 - serialization.ddl struct dest_j1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289272690 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest_j1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: string + expr: _col9 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,val2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 + name dest_j1 + serialization.ddl struct dest_j1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290518817 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest_j1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Path -> Alias: - pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z] Path -> Partition: - pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 Partition base file name: hr=11 input format: org.apache.hadoop.mapred.TextInputFormat @@ -166,13 +161,13 @@ columns.types string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart name srcpart partition_columns ds/hr serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289268000 + transient_lastDdlTime 1290518805 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -183,13 +178,13 @@ columns.types string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart name srcpart partition_columns ds/hr serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289268000 + transient_lastDdlTime 1290518805 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcpart name: srcpart @@ -201,14 +196,14 @@ Move Operator files: hdfs directory: true - source: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10002 - destination: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10002 + destination: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10000 Stage: Stage-0 Move Operator tables: replace: true - source: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -218,28 +213,28 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 serialization.ddl struct dest_j1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289272690 + transient_lastDdlTime 1290518817 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest_j1 - tmp directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10001 + tmp directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10001 Stage: Stage-2 Stats-Aggr Operator - Stats Aggregation Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10000/ + Stats Aggregation Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10000/ Stage: Stage-3 Map Reduce Alias -> Map Operator Tree: - pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10002 File Output Operator compressed: false GlobalTableId: 0 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10000 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10000 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -250,12 +245,12 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 serialization.ddl struct dest_j1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289272690 + transient_lastDdlTime 1290518817 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest_j1 TotalFiles: 1 @@ -263,9 +258,9 @@ MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10002 [pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10002] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10002 [pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10002] Path -> Partition: - pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-10_467_2293155916770670069/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-26-57_702_8677388699150882101/-ext-10002 Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat @@ -276,12 +271,12 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 serialization.ddl struct dest_j1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289272690 + transient_lastDdlTime 1290518817 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -292,12 +287,12 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 serialization.ddl struct dest_j1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289272690 + transient_lastDdlTime 1290518817 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest_j1 name: dest_j1 @@ -329,11 +324,11 @@ PREHOOK: query: select * from dest_j1 x order by x.key PREHOOK: type: QUERY PREHOOK: Input: default@dest_j1 -PREHOOK: Output: file:/tmp/liyintang/hive_2010-11-08_19-18-20_736_2131780804835036352/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-27-11_670_4494790440961342296/-mr-10000 POSTHOOK: query: select * from dest_j1 x order by x.key POSTHOOK: type: QUERY POSTHOOK: Input: default@dest_j1 -POSTHOOK: Output: file:/tmp/liyintang/hive_2010-11-08_19-18-20_736_2131780804835036352/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-27-11_670_4494790440961342296/-mr-10000 POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: dest_j1.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: dest_j1.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] @@ -598,57 +593,52 @@ expr: _col11 type: string outputColumnNames: _col0, _col5, _col9, _col10, _col11 - Filter Operator - isSamplingPred: false - predicate: - expr: ((_col10 = '2008-04-08') and (_col11 = 11)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col9 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,val2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 - name dest_j1 - numFiles 1 - numPartitions 0 - numRows 107 - serialization.ddl struct dest_j1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 - transient_lastDdlTime 1289272700 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest_j1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: string + expr: _col9 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,val2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 + name dest_j1 + numFiles 1 + numPartitions 0 + numRows 107 + serialization.ddl struct dest_j1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 2125 + transient_lastDdlTime 1290518831 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest_j1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Path -> Alias: - pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z] Path -> Partition: - pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 Partition base file name: hr=11 input format: org.apache.hadoop.mapred.TextInputFormat @@ -662,13 +652,13 @@ columns.types string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart name srcpart partition_columns ds/hr serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289268000 + transient_lastDdlTime 1290518805 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -679,13 +669,13 @@ columns.types string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/srcpart + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpart name srcpart partition_columns ds/hr serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289268000 + transient_lastDdlTime 1290518805 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcpart name: srcpart @@ -697,14 +687,14 @@ Move Operator files: hdfs directory: true - source: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10002 - destination: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10002 + destination: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10000 Stage: Stage-0 Move Operator tables: replace: true - source: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -714,7 +704,7 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 numFiles 1 numPartitions 0 @@ -723,23 +713,23 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe totalSize 2125 - transient_lastDdlTime 1289272700 + transient_lastDdlTime 1290518831 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest_j1 - tmp directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10001 + tmp directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10001 Stage: Stage-2 Stats-Aggr Operator - Stats Aggregation Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10000/ + Stats Aggregation Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10000/ Stage: Stage-3 Map Reduce Alias -> Map Operator Tree: - pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10002 File Output Operator compressed: false GlobalTableId: 0 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10000 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10000 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -750,7 +740,7 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 numFiles 1 numPartitions 0 @@ -759,7 +749,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe totalSize 2125 - transient_lastDdlTime 1289272700 + transient_lastDdlTime 1290518831 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest_j1 TotalFiles: 1 @@ -767,9 +757,9 @@ MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10002 [pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10002] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10002 [pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10002] Path -> Partition: - pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_19-18-37_907_7726137335023848601/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-27-34_283_8256365155878225045/-ext-10002 Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat @@ -780,7 +770,7 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 numFiles 1 numPartitions 0 @@ -789,7 +779,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe totalSize 2125 - transient_lastDdlTime 1289272700 + transient_lastDdlTime 1290518831 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -800,7 +790,7 @@ columns.types string:string:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dest_j1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest_j1 name dest_j1 numFiles 1 numPartitions 0 @@ -809,12 +799,11 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe totalSize 2125 - transient_lastDdlTime 1289272700 + transient_lastDdlTime 1290518831 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest_j1 name: dest_j1 - PREHOOK: query: INSERT OVERWRITE TABLE dest_j1 SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value FROM src1_copy x JOIN src_copy y ON (x.key = y.key) @@ -848,11 +837,11 @@ PREHOOK: query: select * from dest_j1 x order by x.key PREHOOK: type: QUERY PREHOOK: Input: default@dest_j1 -PREHOOK: Output: file:/tmp/liyintang/hive_2010-11-08_19-18-49_147_7316712558387738462/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-27-49_503_1806503783455700687/-mr-10000 POSTHOOK: query: select * from dest_j1 x order by x.key POSTHOOK: type: QUERY POSTHOOK: Input: default@dest_j1 -POSTHOOK: Output: file:/tmp/liyintang/hive_2010-11-08_19-18-49_147_7316712558387738462/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-27-49_503_1806503783455700687/-mr-10000 POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1_copy)x.FieldSchema(name:key, type:string, comment:null), ] POSTHOOK: Lineage: dest_j1.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] Index: ql/src/test/results/clientpositive/lateral_view_ppd.q.out =================================================================== --- ql/src/test/results/clientpositive/lateral_view_ppd.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/lateral_view_ppd.q.out (working copy) @@ -34,23 +34,19 @@ expr: _col2 type: int outputColumnNames: _col0, _col1, _col2 - Filter Operator - predicate: - expr: (_col0 = '0') - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Select Operator expressions: expr: array(1,2,3) @@ -69,23 +65,19 @@ expr: _col2 type: int outputColumnNames: _col0, _col1, _col2 - Filter Operator - predicate: - expr: (_col0 = '0') - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -136,19 +128,19 @@ SELECT * : (no compute) Lateral View Join Operator outputColumnNames: _col0, _col1, _col2 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1, _col2 - Filter Operator - predicate: - expr: ((_col0 = '0') and (_col2 = 1)) - type: boolean + Filter Operator + predicate: + expr: ((_col0 = '0') and (_col2 = 1)) + type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1, _col2 Select Operator expressions: expr: _col1 @@ -171,19 +163,19 @@ function name: explode Lateral View Join Operator outputColumnNames: _col0, _col1, _col2 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1, _col2 - Filter Operator - predicate: - expr: ((_col0 = '0') and (_col2 = 1)) - type: boolean + Filter Operator + predicate: + expr: ((_col0 = '0') and (_col2 = 1)) + type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1, _col2 Select Operator expressions: expr: _col1 @@ -248,24 +240,20 @@ expr: _col3 type: string outputColumnNames: _col1, _col4, _col2, _col3 - Filter Operator - predicate: - expr: ((_col2 = '2008-04-08') and (_col3 = '12')) - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col4 - type: int - outputColumnNames: _col0, _col1 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col4 + type: int + outputColumnNames: _col0, _col1 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Select Operator expressions: expr: array(1,2,3) @@ -286,24 +274,20 @@ expr: _col3 type: string outputColumnNames: _col1, _col4, _col2, _col3 - Filter Operator - predicate: - expr: ((_col2 = '2008-04-08') and (_col3 = '12')) - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col4 - type: int - outputColumnNames: _col0, _col1 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col4 + type: int + outputColumnNames: _col0, _col1 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -371,23 +355,19 @@ expr: _col0 type: string outputColumnNames: _col1, _col2, _col0 - Filter Operator - predicate: - expr: (_col0 = '0') - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Select Operator expressions: expr: array(1,2,3) @@ -406,23 +386,19 @@ expr: _col0 type: string outputColumnNames: _col1, _col2, _col0 - Filter Operator - predicate: - expr: (_col0 = '0') - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Select Operator expressions: expr: array(1,2,3) @@ -446,23 +422,19 @@ expr: _col0 type: string outputColumnNames: _col1, _col2, _col0 - Filter Operator - predicate: - expr: (_col0 = '0') - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Select Operator expressions: expr: array(1,2,3) @@ -481,23 +453,19 @@ expr: _col0 type: string outputColumnNames: _col1, _col2, _col0 - Filter Operator - predicate: - expr: (_col0 = '0') - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: int - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: int + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/load_dyn_part10.q.out =================================================================== --- ql/src/test/results/clientpositive/load_dyn_part10.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/load_dyn_part10.q.out (working copy) @@ -44,27 +44,23 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds > '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part10 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part10 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/load_dyn_part13.q.out =================================================================== --- ql/src/test/results/clientpositive/load_dyn_part13.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/load_dyn_part13.q.out (working copy) @@ -62,37 +62,33 @@ predicate: expr: (key < 20) type: boolean - Filter Operator - predicate: - expr: (key < 20) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: '22' - type: string - outputColumnNames: _col0, _col1, _col2 - Union - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part13 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: '22' + type: string + outputColumnNames: _col0, _col1, _col2 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part13 null-subquery2:s-subquery2:src TableScan alias: src @@ -100,37 +96,33 @@ predicate: expr: ((key > 20) and (key < 40)) type: boolean - Filter Operator - predicate: - expr: ((key > 20) and (key < 40)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: '33' - type: string - outputColumnNames: _col0, _col1, _col2 - Union - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part13 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: '33' + type: string + outputColumnNames: _col0, _col1, _col2 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part13 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/load_dyn_part2.q.out =================================================================== --- ql/src/test/results/clientpositive/load_dyn_part2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/load_dyn_part2.q.out (working copy) @@ -38,32 +38,28 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds is not null and hr is not null) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + sort order: + Map-reduce partition columns: + expr: _col0 type: string - expr: value + tag: -1 + value expressions: + expr: _col0 type: string - expr: hr + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - sort order: - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Reduce Operator Tree: Extract File Output Operator Index: ql/src/test/results/clientpositive/load_dyn_part3.q.out =================================================================== --- ql/src/test/results/clientpositive/load_dyn_part3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/load_dyn_part3.q.out (working copy) @@ -42,29 +42,25 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds is not null and hr is not null) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part3 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part3 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/load_dyn_part4.q.out =================================================================== --- ql/src/test/results/clientpositive/load_dyn_part4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/load_dyn_part4.q.out (working copy) @@ -54,29 +54,25 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds is not null and hr is not null) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part4 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part4 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/load_dyn_part9.q.out =================================================================== --- ql/src/test/results/clientpositive/load_dyn_part9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/load_dyn_part9.q.out (working copy) @@ -44,29 +44,25 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds <= '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part9 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part9 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/louter_join_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/louter_join_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/louter_join_ppr.q.out (working copy) @@ -54,25 +54,20 @@ TableScan alias: b GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string Needs Tagging: true Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/src [a] @@ -209,7 +204,7 @@ Filter Operator isSamplingPred: false predicate: - expr: ((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) + expr: ((_col4 > 15) and (_col4 < 25)) type: boolean Select Operator expressions: @@ -575,7 +570,7 @@ Filter Operator isSamplingPred: false predicate: - expr: ((((_col0 > 10) and (_col0 < 20)) and (_col6 > 15)) and (_col6 < 25)) + expr: ((_col6 > 15) and (_col6 < 25)) type: boolean Select Operator expressions: @@ -942,7 +937,7 @@ Filter Operator isSamplingPred: false predicate: - expr: (((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) and (_col6 = '2008-04-08')) + expr: (((_col4 > 15) and (_col4 < 25)) and (_col6 = '2008-04-08')) type: boolean Select Operator expressions: @@ -1225,7 +1220,7 @@ Filter Operator isSamplingPred: false predicate: - expr: (((((_col0 > 10) and (_col0 < 20)) and (_col6 > 15)) and (_col6 < 25)) and (_col2 = '2008-04-08')) + expr: ((_col6 > 15) and (_col6 < 25)) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/mapjoin_subquery.q.out =================================================================== --- ql/src/test/results/clientpositive/mapjoin_subquery.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/mapjoin_subquery.q.out (working copy) @@ -46,23 +46,15 @@ z TableScan alias: z - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - predicate: - expr: (hr = 11) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[key]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce @@ -310,23 +302,15 @@ z TableScan alias: z - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Filter Operator - predicate: - expr: (hr = 11) - type: boolean - HashTable Sink Operator - condition expressions: - 0 {_col0} - 1 {value} - handleSkewJoin: false - keys: - 0 [Column[_col0]] - 1 [Column[key]] - Position of Big Table: 0 + HashTable Sink Operator + condition expressions: + 0 {_col0} + 1 {value} + handleSkewJoin: false + keys: + 0 [Column[_col0]] + 1 [Column[key]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce Index: ql/src/test/results/clientpositive/merge3.q.out =================================================================== --- ql/src/test/results/clientpositive/merge3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/merge3.q.out (working copy) @@ -2267,47 +2267,42 @@ TableScan alias: merge_src_part GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: ds is not null - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/njain/hive1/hive1/build/ql/scratchdir/hive_2010-11-10_21-42-10_192_7350266745211428565/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/njain/hive1/hive1/build/ql/scratchdir/hive_2010-11-10_21-42-10_192_7350266745211428565/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/merge_src_part2 - name merge_src_part2 - partition_columns ds - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289454130 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: merge_src_part2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-22_02-31-29_748_6160879101577377111/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-22_02-31-29_748_6160879101577377111/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/merge_src_part2 + name merge_src_part2 + partition_columns ds + serialization.ddl struct merge_src_part2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290421889 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: merge_src_part2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/merge_src_part/ds=2008-04-08 [merge_src_part] @@ -4645,33 +4640,28 @@ TableScan alias: merge_src_part GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: ds is not null - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + sort order: + Map-reduce partition columns: + expr: _col2 type: string - expr: value + tag: -1 + value expressions: + expr: _col0 type: string - expr: ds + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - sort order: - Map-reduce partition columns: - expr: _col2 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/merge_src_part/ds=2008-04-08 [s:merge_src_part] @@ -4926,6 +4916,7 @@ name: merge_src_part2 name: merge_src_part2 + PREHOOK: query: from (select * from merge_src_part where ds is not null distribute by ds) s insert overwrite table merge_src_part2 partition(ds) select key, value, ds Index: ql/src/test/results/clientpositive/merge4.q.out =================================================================== --- ql/src/test/results/clientpositive/merge4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/merge4.q.out (working copy) @@ -27,27 +27,23 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part Stage: Stage-5 Conditional Operator @@ -1144,25 +1140,21 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_part + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: nzhang_part Stage: Stage-5 Conditional Operator @@ -2774,25 +2766,21 @@ null-subquery1:s-subquery1:srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce Index: ql/src/test/results/clientpositive/merge_dynamic_partition.q.out =================================================================== --- ql/src/test/results/clientpositive/merge_dynamic_partition.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/merge_dynamic_partition.q.out (working copy) @@ -49,27 +49,23 @@ srcpart_merge_dp TableScan alias: srcpart_merge_dp - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: merge_dynamic_part + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: merge_dynamic_part Stage: Stage-0 Move Operator @@ -654,25 +650,21 @@ srcpart_merge_dp TableScan alias: srcpart_merge_dp - Filter Operator - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: merge_dynamic_part + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: merge_dynamic_part Stage: Stage-5 Conditional Operator @@ -1287,29 +1279,25 @@ srcpart_merge_dp TableScan alias: srcpart_merge_dp - Filter Operator - predicate: - expr: ((ds = '2008-04-08') and (hr = 11)) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: merge_dynamic_part + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: merge_dynamic_part Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/no_hooks.q.out =================================================================== --- ql/src/test/results/clientpositive/no_hooks.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/no_hooks.q.out (working copy) @@ -52,27 +52,23 @@ 1 {VALUE._col0} {VALUE._col1} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: ((_col0 < 10) and (_col4 < 10)) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce Index: ql/src/test/results/clientpositive/noalias_subq1.q.out =================================================================== --- ql/src/test/results/clientpositive/noalias_subq1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/noalias_subq1.q.out (working copy) @@ -29,21 +29,17 @@ expr: key type: string outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col1 < 100) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/notable_alias1.q.out =================================================================== --- ql/src/test/results/clientpositive/notable_alias1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/notable_alias1.q.out (working copy) @@ -30,36 +30,32 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/notable_alias2.q.out =================================================================== --- ql/src/test/results/clientpositive/notable_alias2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/notable_alias2.q.out (working copy) @@ -30,36 +30,32 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/nullgroup.q.out =================================================================== --- ql/src/test/results/clientpositive/nullgroup.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/nullgroup.q.out (working copy) @@ -22,23 +22,19 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: bigint + Select Operator + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -96,23 +92,19 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: _col0 - type: bigint + Select Operator + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + mode: hash + outputColumnNames: _col0 + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: _col0 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -171,20 +163,16 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - Reduce Output Operator - sort order: - Map-reduce partition columns: - expr: rand() - type: double - tag: -1 - value expressions: - expr: 1 - type: int + Select Operator + Reduce Output Operator + sort order: + Map-reduce partition columns: + expr: rand() + type: double + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: @@ -266,17 +254,13 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - Reduce Output Operator - sort order: - tag: -1 - value expressions: - expr: 1 - type: int + Select Operator + Reduce Output Operator + sort order: + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/nullgroup2.q.out =================================================================== --- ql/src/test/results/clientpositive/nullgroup2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/nullgroup2.q.out (working copy) @@ -23,36 +23,32 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: rand() - type: double - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: rand() + type: double + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -147,36 +143,32 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: expr: key type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -239,27 +231,23 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Reduce Output Operator + key expressions: expr: key type: string - outputColumnNames: key - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: rand() - type: double - tag: -1 - value expressions: - expr: 1 - type: int + sort order: + + Map-reduce partition columns: + expr: rand() + type: double + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: @@ -354,27 +342,23 @@ predicate: expr: (key > 9999) type: boolean - Filter Operator - predicate: - expr: (key > 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: string + outputColumnNames: key + Reduce Output Operator + key expressions: expr: key type: string - outputColumnNames: key - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: -1 - value expressions: - expr: 1 - type: int + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/nullgroup4.q.out =================================================================== --- ql/src/test/results/clientpositive/nullgroup4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/nullgroup4.q.out (working copy) @@ -23,39 +23,35 @@ predicate: expr: (key = 9999) type: boolean - Filter Operator - predicate: - expr: (key = 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Group By Operator + aggregations: + expr: count(1) + expr: count(DISTINCT value) + bucketGroup: false + keys: expr: value type: string - outputColumnNames: value - Group By Operator - aggregations: - expr: count(1) - expr: count(DISTINCT value) - bucketGroup: false - keys: - expr: value + mode: hash + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint - expr: _col2 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint + expr: _col2 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -143,36 +139,32 @@ predicate: expr: (key = 9999) type: boolean - Filter Operator - predicate: - expr: (key = 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Group By Operator + aggregations: + expr: count(1) + expr: count(DISTINCT value) + bucketGroup: false + keys: expr: value type: string - outputColumnNames: value - Group By Operator - aggregations: - expr: count(1) - expr: count(DISTINCT value) - bucketGroup: false - keys: - expr: value + mode: hash + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col1 - type: bigint - expr: _col2 - type: bigint + sort order: + + tag: -1 + value expressions: + expr: _col1 + type: bigint + expr: _col2 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -234,27 +226,23 @@ predicate: expr: (key = 9999) type: boolean - Filter Operator - predicate: - expr: (key = 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Reduce Output Operator + key expressions: expr: value type: string - outputColumnNames: value - Reduce Output Operator - key expressions: - expr: value - type: string - sort order: + - Map-reduce partition columns: - expr: value - type: string - tag: -1 - value expressions: - expr: 1 - type: int + sort order: + + Map-reduce partition columns: + expr: value + type: string + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: @@ -342,24 +330,20 @@ predicate: expr: (key = 9999) type: boolean - Filter Operator - predicate: - expr: (key = 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Reduce Output Operator + key expressions: expr: value type: string - outputColumnNames: value - Reduce Output Operator - key expressions: - expr: value - type: string - sort order: + - tag: -1 - value expressions: - expr: 1 - type: int + sort order: + + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/nullgroup4_multi_distinct.q.out =================================================================== --- ql/src/test/results/clientpositive/nullgroup4_multi_distinct.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/nullgroup4_multi_distinct.q.out (working copy) @@ -22,43 +22,39 @@ predicate: expr: (key = 9999) type: boolean - Filter Operator - predicate: - expr: (key = 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Group By Operator + aggregations: + expr: count(1) + expr: count(DISTINCT value) + expr: count(DISTINCT substr(value, 5)) + bucketGroup: false + keys: expr: value type: string - outputColumnNames: value - Group By Operator - aggregations: - expr: count(1) - expr: count(DISTINCT value) - expr: count(DISTINCT substr(value, 5)) - bucketGroup: false - keys: - expr: value + expr: substr(value, 5) + type: string + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: substr(value, 5) + expr: _col1 type: string - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col2 - type: bigint - expr: _col3 - type: bigint - expr: _col4 - type: bigint + sort order: ++ + tag: -1 + value expressions: + expr: _col2 + type: bigint + expr: _col3 + type: bigint + expr: _col4 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -122,26 +118,22 @@ predicate: expr: (key = 9999) type: boolean - Filter Operator - predicate: - expr: (key = 9999) - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + outputColumnNames: value + Reduce Output Operator + key expressions: expr: value type: string - outputColumnNames: value - Reduce Output Operator - key expressions: - expr: value - type: string - expr: substr(value, 5) - type: string - sort order: ++ - tag: -1 - value expressions: - expr: 1 - type: int + expr: substr(value, 5) + type: string + sort order: ++ + tag: -1 + value expressions: + expr: 1 + type: int Reduce Operator Tree: Group By Operator aggregations: Index: ql/src/test/results/clientpositive/nullgroup5.q.out =================================================================== --- ql/src/test/results/clientpositive/nullgroup5.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/nullgroup5.q.out (working copy) @@ -52,38 +52,6 @@ predicate: expr: (ds = '2009-04-05') type: boolean - Filter Operator - predicate: - expr: (ds = '2009-04-05') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - null-subquery2:u-subquery2:y - TableScan - alias: y - Filter Operator - predicate: - expr: (ds = '2009-04-09') - type: boolean Select Operator expressions: expr: key @@ -105,6 +73,30 @@ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + null-subquery2:u-subquery2:y + TableScan + alias: y + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/outer_join_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/outer_join_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/outer_join_ppr.q.out (working copy) @@ -288,7 +288,7 @@ Filter Operator isSamplingPred: false predicate: - expr: ((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) + expr: (((_col4 > 15) and (_col4 < 25)) and ((_col0 > 10) and (_col0 < 20))) type: boolean Select Operator expressions: @@ -650,7 +650,7 @@ Filter Operator isSamplingPred: false predicate: - expr: (((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) and (_col6 = '2008-04-08')) + expr: ((((_col4 > 15) and (_col4 < 25)) and (_col6 = '2008-04-08')) and ((_col0 > 10) and (_col0 < 20))) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/pcr.q.out =================================================================== --- ql/src/test/results/clientpositive/pcr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/pcr.q.out (working copy) @@ -81,35 +81,30 @@ predicate: expr: (key < 5) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds <= '2000-04-09') and (key < 5)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col2 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col2 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -291,29 +286,24 @@ predicate: expr: ((ds <= '2000-04-09') or (key < 5)) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds <= '2000-04-09') or (key < 5)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -581,35 +571,30 @@ predicate: expr: ((key < 5) and (value <> 'val_2')) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds <= '2000-04-09') and (key < 5)) and (value <> 'val_2')) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col2 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col2 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -793,35 +778,30 @@ predicate: expr: (((ds < '2000-04-09') and (key < 5)) or ((ds > '2000-04-09') and (value = 'val_5'))) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds < '2000-04-09') and (key < 5)) or ((ds > '2000-04-09') and (value = 'val_5'))) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col2 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col2 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -1007,35 +987,30 @@ predicate: expr: (((ds < '2000-04-10') and (key < 5)) or ((ds > '2000-04-08') and (value = 'val_5'))) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds < '2000-04-10') and (key < 5)) or ((ds > '2000-04-08') and (value = 'val_5'))) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col2 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col2 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -1280,35 +1255,30 @@ predicate: expr: (((ds < '2000-04-10') or (key < 5)) and ((ds > '2000-04-08') or (value = 'val_5'))) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds < '2000-04-10') or (key < 5)) and ((ds > '2000-04-08') or (value = 'val_5'))) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col2 type: string - expr: ds + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col2 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -1561,31 +1531,26 @@ predicate: expr: (key = 14) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds = '2000-04-08') or (ds = '2000-04-09')) and (key = 14)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -1752,31 +1717,26 @@ TableScan alias: pcr_t1 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '2000-04-08') or (ds = '2000-04-09')) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -1983,31 +1943,26 @@ TableScan alias: pcr_t1 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds >= '2000-04-08') or (ds < '2000-04-10')) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -2290,37 +2245,32 @@ predicate: expr: (((ds = '2000-04-08') and (key = 1)) or ((ds = '2000-04-09') and (key = 2))) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds = '2000-04-08') and (key = 1)) or ((ds = '2000-04-09') and (key = 2))) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - expr: ds + expr: _col2 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string - sort order: +++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + sort order: +++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -2489,52 +2439,42 @@ TableScan alias: t1 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2000-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: int - sort order: + - Map-reduce partition columns: - expr: key - type: int - tag: 0 - value expressions: - expr: key - type: int - expr: value - type: string - expr: ds - type: string + Reduce Output Operator + key expressions: + expr: key + type: int + sort order: + + Map-reduce partition columns: + expr: key + type: int + tag: 0 + value expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string t2 TableScan alias: t2 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2000-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: int - sort order: + - Map-reduce partition columns: - expr: key - type: int - tag: 1 - value expressions: - expr: key - type: int - expr: value - type: string - expr: ds - type: string + Reduce Output Operator + key expressions: + expr: key + type: int + sort order: + + Map-reduce partition columns: + expr: key + type: int + tag: 1 + value expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string Needs Tagging: true Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [t2, t1] @@ -2772,52 +2712,42 @@ TableScan alias: t1 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2000-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: int - sort order: + - Map-reduce partition columns: - expr: key - type: int - tag: 0 - value expressions: - expr: key - type: int - expr: value - type: string - expr: ds - type: string + Reduce Output Operator + key expressions: + expr: key + type: int + sort order: + + Map-reduce partition columns: + expr: key + type: int + tag: 0 + value expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string t2 TableScan alias: t2 GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2000-04-09') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: int - sort order: + - Map-reduce partition columns: - expr: key - type: int - tag: 1 - value expressions: - expr: key - type: int - expr: value - type: string - expr: ds - type: string + Reduce Output Operator + key expressions: + expr: key + type: int + sort order: + + Map-reduce partition columns: + expr: key + type: int + tag: 1 + value expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string Needs Tagging: true Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [t1] @@ -3128,37 +3058,32 @@ predicate: expr: (((ds > '2000-04-08') and (ds < '2000-04-11')) or (key = 2)) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds > '2000-04-08') and (ds < '2000-04-11')) or (((ds >= '2000-04-08') and (ds <= '2000-04-11')) and (key = 2))) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - expr: ds + expr: _col2 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string - sort order: +++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + sort order: +++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -3478,37 +3403,32 @@ predicate: expr: ((ds > '2000-04-08') or ((ds <= '2000-04-09') and (key = 2))) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds > '2000-04-08') and (ds < '2000-04-11')) or ((ds <= '2000-04-09') and (key = 2))) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + expr: ds + type: string + outputColumnNames: _col0, _col1, _col2 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - expr: ds + expr: _col2 type: string - outputColumnNames: _col0, _col1, _col2 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string - sort order: +++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col2 - type: string + sort order: +++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/pcr_t1/ds=2000-04-08 [pcr_t1] @@ -4671,29 +4591,24 @@ TableScan alias: srcpart GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '2008-04-08') and (hr = 11)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + tag: -1 + value expressions: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + expr: _col1 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [srcpart] @@ -4826,41 +4741,36 @@ predicate: expr: (key = 11) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds = '2008-04-08') and ((hr = '11') or (hr = '12'))) and (key = 11)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col2 type: string - expr: ds + expr: _col3 type: string - expr: hr + sort order: +++ + tag: -1 + value expressions: + expr: _col0 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - sort order: +++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [srcpart] @@ -5040,41 +4950,36 @@ predicate: expr: (key = 11) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((hr = '11') and (key = 11)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + expr: _col2 type: string - expr: ds + expr: _col3 type: string - expr: hr + sort order: +++ + tag: -1 + value expressions: + expr: _col0 type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - sort order: +++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/open-source-hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [srcpart] Index: ql/src/test/results/clientpositive/ppd1.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd1.q.out (working copy) @@ -22,21 +22,17 @@ predicate: expr: (key > '2') type: boolean - Filter Operator - predicate: - expr: (key > '2') - type: boolean - Select Operator - expressions: - expr: key - type: string - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/ppd_clusterby.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_clusterby.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_clusterby.q.out (working copy) @@ -22,31 +22,27 @@ predicate: expr: (key = 10) type: boolean - Filter Operator - predicate: - expr: (key = 10) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract File Output Operator @@ -133,25 +129,21 @@ 1 {VALUE._col0} handleSkewJoin: false outputColumnNames: _col0, _col1, _col4 - Filter Operator - predicate: - expr: (_col0 = 20) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col4 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col4 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce Index: ql/src/test/results/clientpositive/ppd_gby.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_gby.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_gby.q.out (working copy) @@ -28,38 +28,34 @@ predicate: expr: ((value > 'val_10') and (value > 'val_200')) type: boolean - Filter Operator - predicate: - expr: (value > 'val_10') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + expr: key + type: string + outputColumnNames: value, key + Group By Operator + aggregations: + expr: count(key) + bucketGroup: false + keys: expr: value type: string - expr: key - type: string - outputColumnNames: value, key - Group By Operator - aggregations: - expr: count(key) - bucketGroup: false - keys: - expr: value + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -70,17 +66,17 @@ type: string mode: mergepartial outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: bigint - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: ((_col0 > 'val_200') and ((_col1 > 30) or (_col0 < 'val_400'))) - type: boolean + Filter Operator + predicate: + expr: ((_col0 > 'val_200') and ((_col1 > 30) or (_col0 < 'val_400'))) + type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: bigint + outputColumnNames: _col0, _col1 Select Operator expressions: expr: _col0 Index: ql/src/test/results/clientpositive/ppd_gby2.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_gby2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_gby2.q.out (working copy) @@ -31,38 +31,34 @@ predicate: expr: ((value > 'val_10') and (value > 'val_200')) type: boolean - Filter Operator - predicate: - expr: (value > 'val_10') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: value + type: string + expr: key + type: string + outputColumnNames: value, key + Group By Operator + aggregations: + expr: count(key) + bucketGroup: false + keys: expr: value type: string - expr: key - type: string - outputColumnNames: value, key - Group By Operator - aggregations: - expr: count(key) - bucketGroup: false - keys: - expr: value + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Reduce Operator Tree: Group By Operator aggregations: @@ -73,17 +69,17 @@ type: string mode: mergepartial outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: bigint - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: ((_col0 > 'val_200') and ((_col1 > 30) or (_col0 < 'val_400'))) - type: boolean + Filter Operator + predicate: + expr: ((_col0 > 'val_200') and ((_col1 > 30) or (_col0 < 'val_400'))) + type: boolean + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: bigint + outputColumnNames: _col0, _col1 Select Operator expressions: expr: _col0 Index: ql/src/test/results/clientpositive/ppd_gby_join.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_gby_join.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_gby_join.q.out (working copy) @@ -37,35 +37,27 @@ predicate: expr: (((key > '1') and (key < '400')) and ((key > '20') and ((value < 'val_50') or (key > '2')))) type: boolean - Filter Operator - predicate: - expr: (key > '1') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 < '400') - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string src2:src TableScan alias: src @@ -73,27 +65,23 @@ predicate: expr: ((key > '2') and (key <> '4')) type: boolean - Filter Operator - predicate: - expr: (key > '2') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + outputColumnNames: _col0 + Reduce Output Operator + key expressions: + expr: _col0 type: string - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 1 + value expressions: + expr: _col0 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/ppd_join.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_join.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_join.q.out (working copy) @@ -34,35 +34,27 @@ predicate: expr: (((key > '1') and (key < '400')) and ((key > '20') and ((value < 'val_50') or (key > '2')))) type: boolean - Filter Operator - predicate: - expr: (key > '1') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 < '400') - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string src2:src TableScan alias: src @@ -70,31 +62,27 @@ predicate: expr: ((key > '2') and (key <> '4')) type: boolean - Filter Operator - predicate: - expr: (key > '2') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/ppd_join2.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_join2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_join2.q.out (working copy) @@ -41,35 +41,27 @@ predicate: expr: (((key <> '302') and (key < '400')) and ((key <> '311') and ((value <> 'val_50') or (key > '1')))) type: boolean - Filter Operator - predicate: - expr: (key <> '302') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 < '400') - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string src2:src TableScan alias: src @@ -77,31 +69,27 @@ predicate: expr: ((key <> '305') and (key <> '14')) type: boolean - Filter Operator - predicate: - expr: (key <> '305') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: @@ -147,29 +135,25 @@ predicate: expr: ((key <> '306') and (sqrt(key) <> 13)) type: boolean - Filter Operator - predicate: - expr: (key <> '306') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col1 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col1 - type: string - sort order: + - Map-reduce partition columns: - expr: _col1 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/ppd_join3.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_join3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_join3.q.out (working copy) @@ -40,35 +40,27 @@ predicate: expr: (((key <> '11') and (key < '400')) and ((key > '0') and ((value <> 'val_500') or (key > '1')))) type: boolean - Filter Operator - predicate: - expr: (key <> '11') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 < '400') - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string src2:src TableScan alias: src @@ -76,31 +68,27 @@ predicate: expr: ((key <> '12') and (key <> '4')) type: boolean - Filter Operator - predicate: - expr: (key <> '12') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string src3:src TableScan alias: src @@ -108,27 +96,23 @@ predicate: expr: ((key <> '13') and (key <> '1')) type: boolean - Filter Operator - predicate: - expr: (key <> '13') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + outputColumnNames: _col0 + Reduce Output Operator + key expressions: + expr: _col0 type: string - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 - value expressions: - expr: _col0 - type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 2 + value expressions: + expr: _col0 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/ppd_outer_join1.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_outer_join1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_outer_join1.q.out (working copy) @@ -76,7 +76,7 @@ outputColumnNames: _col0, _col1, _col4, _col5 Filter Operator predicate: - expr: ((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) + expr: ((_col4 > 15) and (_col4 < 25)) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/ppd_outer_join2.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_outer_join2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_outer_join2.q.out (working copy) @@ -76,7 +76,7 @@ outputColumnNames: _col0, _col1, _col4, _col5 Filter Operator predicate: - expr: ((((_col0 > '10') and (_col0 < '20')) and (_col4 > '15')) and (_col4 < '25')) + expr: ((_col0 > '10') and (_col0 < '20')) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/ppd_outer_join3.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_outer_join3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_outer_join3.q.out (working copy) @@ -72,7 +72,7 @@ outputColumnNames: _col0, _col1, _col4, _col5 Filter Operator predicate: - expr: ((((_col0 > '10') and (_col0 < '20')) and (_col4 > '15')) and (_col4 < '25')) + expr: (((_col4 > '15') and (_col4 < '25')) and ((_col0 > '10') and (_col0 < '20'))) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/ppd_outer_join4.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_outer_join4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_outer_join4.q.out (working copy) @@ -95,7 +95,7 @@ outputColumnNames: _col0, _col1, _col4, _col5, _col8 Filter Operator predicate: - expr: (((((_col0 > '10') and (_col0 < '20')) and (_col4 > '15')) and (_col4 < '25')) and (sqrt(_col8) <> 13)) + expr: ((((_col4 > '15') and (_col4 < '25')) and (sqrt(_col8) <> 13)) and ((_col0 > '10') and (_col0 < '20'))) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/ppd_random.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_random.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_random.q.out (working copy) @@ -54,29 +54,25 @@ predicate: expr: (key > '2') type: boolean - Filter Operator - predicate: - expr: (key > '2') - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/ppd_transform.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_transform.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_transform.q.out (working copy) @@ -42,39 +42,39 @@ output info: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Filter Operator + predicate: + expr: (_col0 < 100) + type: boolean + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Extract - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/ppd_udf_case.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_udf_case.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_udf_case.q.out (working copy) @@ -91,35 +91,31 @@ 1 {VALUE._col0} {VALUE._col1} {VALUE._col2} {VALUE._col3} handleSkewJoin: false outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7, _col8, _col9 - Filter Operator - predicate: - expr: (((_col2 = '2008-04-08') and (_col8 = '2008-04-08')) and CASE (_col0) WHEN ('27') THEN (true) WHEN ('38') THEN (false) ELSE (null) END) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - expr: _col6 - type: string - expr: _col7 - type: string - expr: _col8 - type: string - expr: _col9 - type: string - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + expr: _col6 + type: string + expr: _col7 + type: string + expr: _col8 + type: string + expr: _col9 + type: string + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-2 Map Reduce Index: ql/src/test/results/clientpositive/ppd_union.q.out =================================================================== --- ql/src/test/results/clientpositive/ppd_union.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/ppd_union.q.out (working copy) @@ -34,35 +34,27 @@ predicate: expr: ((key < '100') and ((key > '4') and (value > 'val_4'))) type: boolean - Filter Operator - predicate: - expr: (key < '100') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Filter Operator - predicate: - expr: ((_col0 > '4') and (_col1 > 'val_4')) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat null-subquery2:unioned_query-subquery2:src TableScan alias: src @@ -70,35 +62,27 @@ predicate: expr: ((key > '150') and ((key > '4') and (value > 'val_4'))) type: boolean - Filter Operator - predicate: - expr: (key > '150') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Filter Operator - predicate: - expr: ((_col0 > '4') and (_col1 > 'val_4')) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/query_result_fileformat.q.out =================================================================== --- ql/src/test/results/clientpositive/query_result_fileformat.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/query_result_fileformat.q.out (working copy) @@ -56,23 +56,19 @@ predicate: expr: (key = 'key1') type: boolean - Filter Operator - predicate: - expr: (key = 'key1') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -136,23 +132,19 @@ predicate: expr: (key = 'key1') type: boolean - Filter Operator - predicate: - expr: (key = 'key1') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/quote1.q.out =================================================================== --- ql/src/test/results/clientpositive/quote1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/quote1.q.out (working copy) @@ -33,32 +33,28 @@ predicate: expr: ((key >= 200) and (key < 300)) type: boolean - Filter Operator - predicate: - expr: ((key >= 200) and (key < 300)) - type: boolean + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: key + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: value - type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator @@ -122,25 +118,21 @@ predicate: expr: (table = '2008-04-08') type: boolean - Filter Operator - predicate: - expr: (table = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: location - type: int - expr: type - type: string - expr: table - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: location + type: int + expr: type + type: string + expr: table + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/rand_partitionpruner3.q.out =================================================================== --- ql/src/test/results/clientpositive/rand_partitionpruner3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/rand_partitionpruner3.q.out (working copy) @@ -143,38 +143,33 @@ predicate: expr: (not ((key > 50) or (key < 10))) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (((ds = '2008-04-08') and (not ((key > 50) or (key < 10)))) and (hr like '%2')) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/sdong/hive_2010-11-03_12-58-29_501_7521924359924125192/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/sdong/hive_2010-11-03_12-58-29_501_7521924359924125192/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_01-21-12_789_7125254108453642823/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_01-21-12_789_7125254108453642823/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1,_col2,_col3 + columns.types string:string:string:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [a] Index: ql/src/test/results/clientpositive/rcfile_null_value.q.out =================================================================== --- ql/src/test/results/clientpositive/rcfile_null_value.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/rcfile_null_value.q.out (working copy) @@ -106,31 +106,27 @@ predicate: expr: ((key > 10) and (key < 20)) type: boolean - Filter Operator - predicate: - expr: ((key > 10) and (key < 20)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 0 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string c:b:src2 TableScan alias: src2 @@ -138,31 +134,27 @@ predicate: expr: ((key > 15) and (key < 25)) type: boolean - Filter Operator - predicate: - expr: ((key > 15) and (key < 25)) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - expr: value + sort order: + + Map-reduce partition columns: + expr: _col0 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + tag: 1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/reduce_deduplicate.q.out =================================================================== --- ql/src/test/results/clientpositive/reduce_deduplicate.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/reduce_deduplicate.q.out (working copy) @@ -258,52 +258,47 @@ predicate: expr: (ds = '2010-03-29') type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2010-03-29') - type: boolean - Select Operator - expressions: - expr: aet - type: string - expr: aes - type: string - outputColumnNames: _col0, _col1 - Transform Operator - command: cat - output info: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 - columns.types string,string,int,string,bigint,string,string - field.delim 9 - serialization.format 9 - Reduce Output Operator - key expressions: - expr: _col1 - type: string - sort order: + - Map-reduce partition columns: - expr: _col1 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: int - expr: _col3 - type: string - expr: _col4 - type: bigint - expr: _col5 - type: string - expr: _col6 - type: string + Select Operator + expressions: + expr: aet + type: string + expr: aes + type: string + outputColumnNames: _col0, _col1 + Transform Operator + command: cat + output info: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 + columns.types string,string,int,string,bigint,string,string + field.delim 9 + serialization.format 9 + Reduce Output Operator + key expressions: + expr: _col1 + type: string + sort order: + + Map-reduce partition columns: + expr: _col1 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: int + expr: _col3 + type: string + expr: _col4 + type: bigint + expr: _col5 + type: string + expr: _col6 + type: string Needs Tagging: false Reduce Operator Tree: Extract Index: ql/src/test/results/clientpositive/regex_col.q.out =================================================================== --- ql/src/test/results/clientpositive/regex_col.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/regex_col.q.out (working copy) @@ -216,27 +216,23 @@ predicate: expr: (key = 103) type: boolean - Filter Operator - predicate: - expr: (key = 103) - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - expr: hr - type: string - expr: ds - type: string - sort order: +++ - Map-reduce partition columns: - expr: key - type: string - expr: hr - type: string - expr: ds - type: string - tag: 0 + Reduce Output Operator + key expressions: + expr: key + type: string + expr: hr + type: string + expr: ds + type: string + sort order: +++ + Map-reduce partition columns: + expr: key + type: string + expr: hr + type: string + expr: ds + type: string + tag: 0 b TableScan alias: b Index: ql/src/test/results/clientpositive/regexp_extract.q.out =================================================================== --- ql/src/test/results/clientpositive/regexp_extract.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/regexp_extract.q.out (working copy) @@ -53,20 +53,25 @@ field.delim 9 serialization.format 9 serialization.last.column.takes.rest true - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Filter Operator + isSamplingPred: false + predicate: + expr: (_col0 < 100) + type: boolean + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/src [tmap:src] @@ -109,34 +114,29 @@ name: src Reduce Operator Tree: Extract - Filter Operator - isSamplingPred: false - predicate: - expr: (_col0 < 100) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: regexp_extract(_col1, 'val_(\d+\t\d+)', 1) - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-22-34_668_2056813096308329710/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-22-34_668_2056813096308329710/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1 - columns.types string:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: string + expr: regexp_extract(_col1, 'val_(\d+\t\d+)', 1) + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_01-24-29_519_2554711952452469298/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_01-24-29_519_2554711952452469298/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1 + columns.types string:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator @@ -302,20 +302,25 @@ field.delim 9 serialization.format 9 serialization.last.column.takes.rest true - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Filter Operator + isSamplingPred: false + predicate: + expr: (_col0 < 100) + type: boolean + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/src [tmap:src] @@ -358,34 +363,29 @@ name: src Reduce Operator Tree: Extract - Filter Operator - isSamplingPred: false - predicate: - expr: (_col0 < 100) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: regexp_extract(_col1, 'val_(\d+\t\d+)') - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-22-38_507_8896713942915944466/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-22-38_507_8896713942915944466/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1 - columns.types string:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: string + expr: regexp_extract(_col1, 'val_(\d+\t\d+)') + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_01-24-37_899_3565519331651273390/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_01-24-37_899_3565519331651273390/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1 + columns.types string:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/router_join_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/router_join_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/router_join_ppr.q.out (working copy) @@ -293,7 +293,7 @@ Filter Operator isSamplingPred: false predicate: - expr: ((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) + expr: ((_col0 > 10) and (_col0 < 20)) type: boolean Select Operator expressions: @@ -401,25 +401,20 @@ TableScan alias: a GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Reduce Output Operator - key expressions: - expr: key - type: string - sort order: + - Map-reduce partition columns: - expr: key - type: string - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + key expressions: + expr: key + type: string + sort order: + + Map-reduce partition columns: + expr: key + type: string + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string b TableScan alias: b @@ -579,7 +574,7 @@ Filter Operator isSamplingPred: false predicate: - expr: ((((_col0 > 10) and (_col0 < 20)) and (_col6 > 15)) and (_col6 < 25)) + expr: ((_col0 > 10) and (_col0 < 20)) type: boolean Select Operator expressions: @@ -858,7 +853,7 @@ Filter Operator isSamplingPred: false predicate: - expr: (((((_col0 > 10) and (_col0 < 20)) and (_col4 > 15)) and (_col4 < 25)) and (_col6 = '2008-04-08')) + expr: ((_col0 > 10) and (_col0 < 20)) type: boolean Select Operator expressions: @@ -1221,7 +1216,7 @@ Filter Operator isSamplingPred: false predicate: - expr: (((((_col0 > 10) and (_col0 < 20)) and (_col6 > 15)) and (_col6 < 25)) and (_col2 = '2008-04-08')) + expr: (((_col0 > 10) and (_col0 < 20)) and (_col2 = '2008-04-08')) type: boolean Select Operator expressions: Index: ql/src/test/results/clientpositive/sample1.q.out =================================================================== --- ql/src/test/results/clientpositive/sample1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample1.q.out (working copy) @@ -39,59 +39,54 @@ predicate: expr: (((hash(rand()) & 2147483647) % 1) = 0) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '2008-04-08') and (hr = '11')) - type: boolean + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 Select Operator expressions: - expr: key + expr: UDFToInteger(_col0) + type: int + expr: _col1 type: string - expr: value + expr: _col2 type: string - expr: ds + expr: _col3 type: string - expr: hr - type: string outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/sdong/www/hive-trunk/build/ql/scratchdir/hive_2010-11-01_15-04-13_003_6538249259772726554/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/sdong/www/hive-trunk/build/ql/scratchdir/hive_2010-11-01_15-04-13_003_6538249259772726554/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value,dt,hr - columns.types int:string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value, string dt, string hr} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1288649052 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_01-25-40_201_2740303071337309249/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_01-25-40_201_2740303071337309249/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value,dt,hr + columns.types int:string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value, string dt, string hr} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289985940 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [s] Index: ql/src/test/results/clientpositive/sample10.q.out =================================================================== --- ql/src/test/results/clientpositive/sample10.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample10.q.out (working copy) @@ -62,56 +62,46 @@ alias: srcpartbucket GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 4) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 4) = 0) - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ds is not null - type: boolean - Select Operator - expressions: - expr: ds + Select Operator + expressions: + expr: ds + type: string + outputColumnNames: ds + Group By Operator + aggregations: + expr: count(1) + bucketGroup: false + keys: + expr: ds + type: string + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: string - outputColumnNames: ds - Group By Operator - aggregations: - expr: count(1) - bucketGroup: false - keys: - expr: ds - type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: -1 + value expressions: + expr: _col1 + type: bigint Needs Tagging: false Path -> Alias: - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11/000000_0 [srcpartbucket] - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12/000000_0 [srcpartbucket] - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11/000000_0 [srcpartbucket] - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12/000000_0 [srcpartbucket] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11 [srcpartbucket] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12 [srcpartbucket] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11 [srcpartbucket] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12 [srcpartbucket] Path -> Partition: - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11/000000_0 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11 Partition - base file name: 000000_0 + base file name: hr=11 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -124,7 +114,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -134,7 +124,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -146,7 +136,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -156,13 +146,13 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: srcpartbucket name: srcpartbucket - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12/000000_0 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12 Partition - base file name: 000000_0 + base file name: hr=12 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -175,7 +165,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -185,7 +175,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -197,7 +187,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -207,13 +197,13 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: srcpartbucket name: srcpartbucket - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11/000000_0 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11 Partition - base file name: 000000_0 + base file name: hr=11 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -226,7 +216,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -236,7 +226,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -248,7 +238,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -258,13 +248,13 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: srcpartbucket name: srcpartbucket - pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12/000000_0 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12 Partition - base file name: 000000_0 + base file name: hr=12 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -277,7 +267,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -287,7 +277,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -299,7 +289,7 @@ columns.types string:string file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat - location pfile:/data/users/njain/hive1/hive1/build/ql/test/data/warehouse/srcpartbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcpartbucket name srcpartbucket numFiles 16 numPartitions 4 @@ -309,7 +299,7 @@ serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe totalSize 2748 - transient_lastDdlTime 1289454630 + transient_lastDdlTime 1290519224 serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: srcpartbucket name: srcpartbucket @@ -333,9 +323,9 @@ File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/njain/hive_2010-11-10_21-50-30_237_6456899979504390679/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-33-44_388_6715985518220677852/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/njain/hive_2010-11-10_21-50-30_237_6456899979504390679/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-33-44_388_6715985518220677852/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -358,13 +348,14 @@ PREHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12 -PREHOOK: Output: file:/tmp/njain/hive_2010-11-10_21-50-31_589_3688459449074450497/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-33-44_819_1118762611602249641/-mr-10000 POSTHOOK: query: select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) where ds is not null group by ds POSTHOOK: type: QUERY POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-33-44_819_1118762611602249641/-mr-10000 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] @@ -381,13 +372,14 @@ PREHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12 -PREHOOK: Output: file:/tmp/njain/hive_2010-11-10_21-50-40_846_194213387409754762/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-33-57_290_4878851869994025804/-mr-10000 POSTHOOK: query: select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 2 on key) where ds is not null group by ds POSTHOOK: type: QUERY POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-33-57_290_4878851869994025804/-mr-10000 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] @@ -404,13 +396,14 @@ PREHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12 -PREHOOK: Output: file:/tmp/njain/hive_2010-11-10_21-50-55_151_7435896620054029141/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-34-11_473_6233665661683805879/-mr-10000 POSTHOOK: query: select * from srcpartbucket where ds is not null POSTHOOK: type: QUERY POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-34-11_473_6233665661683805879/-mr-10000 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] Index: ql/src/test/results/clientpositive/sample2.q.out =================================================================== --- ql/src/test/results/clientpositive/sample2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample2.q.out (working copy) @@ -35,55 +35,50 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510248 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290519453 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition - base file name: srcbucket0.txt + base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -93,12 +88,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519447 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -110,12 +105,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519447 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -127,14 +122,14 @@ Move Operator files: hdfs directory: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10002 - destination: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10002 + destination: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10000 Stage: Stage-0 Move Operator tables: replace: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -144,28 +139,28 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510248 + transient_lastDdlTime 1290519453 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 - tmp directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10001 + tmp directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10001 Stage: Stage-2 Stats-Aggr Operator - Stats Aggregation Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10000/ + Stats Aggregation Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10000/ Stage: Stage-3 Map Reduce Alias -> Map Operator Tree: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10002 File Output Operator compressed: false GlobalTableId: 0 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10000 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10000 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -176,12 +171,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510248 + transient_lastDdlTime 1290519453 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 TotalFiles: 1 @@ -189,9 +184,9 @@ MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10002 [pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10002] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10002 [pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10002] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-08_796_3592086051996573769/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-37-33_168_7009310298289254684/-ext-10002 Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat @@ -202,12 +197,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510248 + transient_lastDdlTime 1290519453 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -218,12 +213,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510248 + transient_lastDdlTime 1290519453 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 name: dest1 @@ -244,11 +239,11 @@ PREHOOK: query: SELECT dest1.* FROM dest1 PREHOOK: type: QUERY PREHOOK: Input: default@dest1 -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-15_725_5572548852641754603/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-37-46_996_2582885276969588759/-mr-10000 POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-15_725_5572548852641754603/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-37-46_996_2582885276969588759/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 474 val_475 Index: ql/src/test/results/clientpositive/sample3.q.out =================================================================== --- ql/src/test/results/clientpositive/sample3.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample3.q.out (working copy) @@ -26,21 +26,17 @@ predicate: expr: (((hash(key) & 2147483647) % 5) = 0) type: boolean - Filter Operator - predicate: - expr: (((hash(key) & 2147483647) % 5) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: int - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: int + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/sample4.q.out =================================================================== --- ql/src/test/results/clientpositive/sample4.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample4.q.out (working copy) @@ -35,55 +35,50 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510262 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290519636 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition - base file name: srcbucket0.txt + base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -93,12 +88,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519629 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -110,12 +105,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519629 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -127,14 +122,14 @@ Move Operator files: hdfs directory: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10002 - destination: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10002 + destination: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10000 Stage: Stage-0 Move Operator tables: replace: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -144,28 +139,28 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510262 + transient_lastDdlTime 1290519636 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 - tmp directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10001 + tmp directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10001 Stage: Stage-2 Stats-Aggr Operator - Stats Aggregation Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10000/ + Stats Aggregation Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10000/ Stage: Stage-3 Map Reduce Alias -> Map Operator Tree: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10002 File Output Operator compressed: false GlobalTableId: 0 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10000 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10000 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -176,12 +171,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510262 + transient_lastDdlTime 1290519636 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 TotalFiles: 1 @@ -189,9 +184,9 @@ MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10002 [pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10002] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10002 [pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10002] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-22_823_2600373229446067176/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-40-36_720_7518621373511334897/-ext-10002 Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat @@ -202,12 +197,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510262 + transient_lastDdlTime 1290519636 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -218,12 +213,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510262 + transient_lastDdlTime 1290519636 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 name: dest1 @@ -244,11 +239,11 @@ PREHOOK: query: SELECT dest1.* FROM dest1 PREHOOK: type: QUERY PREHOOK: Input: default@dest1 -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-28_726_8098117195006938273/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-40-56_148_6748313837250033097/-mr-10000 POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-28_726_8098117195006938273/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-40-56_148_6748313837250033097/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 474 val_475 Index: ql/src/test/results/clientpositive/sample5.q.out =================================================================== --- ql/src/test/results/clientpositive/sample5.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample5.q.out (working copy) @@ -33,48 +33,43 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 5) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 5) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-29_986_2811943230690720353/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-29_986_2811943230690720353/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510269 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-22_02-51-53_792_8607615974872519340/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-22_02-51-53_792_8607615974872519340/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290423113 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket [s] Index: ql/src/test/results/clientpositive/sample6.q.out =================================================================== --- ql/src/test/results/clientpositive/sample6.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample6.q.out (working copy) @@ -33,55 +33,50 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 4) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 4) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510285 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290519881 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition - base file name: srcbucket0.txt + base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -91,12 +86,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -108,12 +103,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -125,14 +120,14 @@ Move Operator files: hdfs directory: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10002 - destination: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10002 + destination: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10000 Stage: Stage-0 Move Operator tables: replace: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -142,28 +137,28 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510285 + transient_lastDdlTime 1290519881 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 - tmp directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10001 + tmp directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10001 Stage: Stage-2 Stats-Aggr Operator - Stats Aggregation Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10000/ + Stats Aggregation Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10000/ Stage: Stage-3 Map Reduce Alias -> Map Operator Tree: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10002 File Output Operator compressed: false GlobalTableId: 0 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10000 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10000 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -174,12 +169,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510285 + transient_lastDdlTime 1290519881 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 TotalFiles: 1 @@ -187,9 +182,9 @@ MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10002 [pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10002] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10002 [pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10002] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-24-46_014_8924466582548738763/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-44-41_962_5733349783644264217/-ext-10002 Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat @@ -200,12 +195,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510285 + transient_lastDdlTime 1290519881 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -216,12 +211,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510285 + transient_lastDdlTime 1290519881 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 name: dest1 @@ -242,11 +237,11 @@ PREHOOK: query: SELECT dest1.* FROM dest1 PREHOOK: type: QUERY PREHOOK: Input: default@dest1 -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-52_602_2662316424952173793/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-44-56_536_3272139135428352120/-mr-10000 POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-52_602_2662316424952173793/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-44-56_536_3272139135428352120/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 468 val_469 @@ -522,42 +517,37 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 4) = 3) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 4) = 3) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket1.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket1.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition - base file name: srcbucket1.txt + base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -567,12 +557,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -584,12 +574,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -598,9 +588,9 @@ File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-24-53_002_533983126018968346/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-44-56_997_9157283476548650979/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-24-53_002_533983126018968346/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-44-56_997_9157283476548650979/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -621,12 +611,12 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-53_250_4539562322091227053/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-44-57_120_2282609152038307645/-mr-10000 POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 4 OUT OF 4 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-53_250_4539562322091227053/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-44-57_120_2282609152038307645/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 3 val_4 @@ -892,42 +882,37 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition - base file name: srcbucket0.txt + base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -937,12 +922,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -954,12 +939,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -968,9 +953,9 @@ File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-24-57_026_2718577770255458785/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-03_958_4413222253884979033/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-24-57_026_2718577770255458785/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-03_958_4413222253884979033/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -991,12 +976,12 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-57_277_4821900235297777819/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-04_108_4466211023967118848/-mr-10000 POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-24-57_277_4821900235297777819/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-04_108_4466211023967118848/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 0 val_0 @@ -1516,40 +1501,35 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 3) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 3) = 0) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat @@ -1561,12 +1541,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1578,12 +1558,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -1592,9 +1572,9 @@ File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-25-01_051_6447605395208317671/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-11_932_8064564472370551552/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-25-01_051_6447605395208317671/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-11_932_8064564472370551552/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -1615,12 +1595,12 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-01_161_3575686286120830874/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-12_066_61352810995258511/-mr-10000 POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 3 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-01_161_3575686286120830874/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-12_066_61352810995258511/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 0 val_0 @@ -1983,40 +1963,35 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 3) = 1) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 3) = 1) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat @@ -2028,12 +2003,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -2045,12 +2020,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290519875 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -2059,9 +2034,9 @@ File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-25-05_752_5484487891016862969/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-20_343_3158771114515930568/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-25-05_752_5484487891016862969/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-20_343_3158771114515930568/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -2082,12 +2057,12 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-05_858_6264519128243618167/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-20_468_7268741501904543933/-mr-10000 POSTHOOK: query: SELECT s.* FROM srcbucket TABLESAMPLE (BUCKET 2 OUT OF 3 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-05_858_6264519128243618167/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-20_468_7268741501904543933/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 1 val_2 @@ -2436,43 +2411,37 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2/srcbucket20.txt [s] - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2/srcbucket22.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2/srcbucket20.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 Partition - base file name: srcbucket20.txt + base file name: srcbucket2 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -2482,12 +2451,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 name srcbucket2 serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504428 + transient_lastDdlTime 1290519877 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -2499,61 +2468,23 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 name srcbucket2 serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504428 + transient_lastDdlTime 1290519877 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket2 name: srcbucket2 - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2/srcbucket22.txt - Partition - base file name: srcbucket22.txt - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count 4 - bucket_field_name key - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2 - name srcbucket2 - serialization.ddl struct srcbucket2 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504428 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count 4 - bucket_field_name key - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2 - name srcbucket2 - serialization.ddl struct srcbucket2 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504428 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: srcbucket2 - name: srcbucket2 Reduce Operator Tree: Extract File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-25-10_451_2535168614473791178/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-29_091_7213447165205917656/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-25-10_451_2535168614473791178/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-29_091_7213447165205917656/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -2574,12 +2505,12 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket2 -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-11_181_8784441563940501492/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-29_242_7529747153649192390/-mr-10000 POSTHOOK: query: SELECT s.* FROM srcbucket2 TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket2 -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-11_181_8784441563940501492/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-29_242_7529747153649192390/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 0 val_0 @@ -2588,26 +2519,62 @@ 2 val_2 4 val_4 8 val_8 +10 val_10 +12 val_12 +12 val_12 +18 val_18 +18 val_18 20 val_20 24 val_24 24 val_24 26 val_26 26 val_26 28 val_28 +30 val_30 +34 val_34 42 val_42 42 val_42 44 val_44 +54 val_54 +58 val_58 +58 val_58 64 val_64 66 val_66 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +78 val_78 80 val_80 82 val_82 84 val_84 84 val_84 86 val_86 +90 val_90 +90 val_90 +90 val_90 +92 val_92 +96 val_96 +98 val_98 +98 val_98 +100 val_100 +100 val_100 +104 val_104 +104 val_104 114 val_114 116 val_116 118 val_118 118 val_118 +120 val_120 +120 val_120 +126 val_126 +128 val_128 +128 val_128 +128 val_128 134 val_134 134 val_134 136 val_136 @@ -2615,11 +2582,19 @@ 138 val_138 138 val_138 138 val_138 +146 val_146 +146 val_146 150 val_150 152 val_152 152 val_152 156 val_156 158 val_158 +160 val_160 +162 val_162 +164 val_164 +164 val_164 +166 val_166 +168 val_168 170 val_170 172 val_172 172 val_172 @@ -2628,6 +2603,8 @@ 176 val_176 176 val_176 178 val_178 +180 val_180 +186 val_186 190 val_190 192 val_192 194 val_194 @@ -2638,18 +2615,38 @@ 208 val_208 208 val_208 208 val_208 +214 val_214 +216 val_216 +216 val_216 +218 val_218 222 val_222 224 val_224 224 val_224 226 val_226 228 val_228 +230 val_230 +230 val_230 +230 val_230 +230 val_230 +230 val_230 +238 val_238 +238 val_238 242 val_242 242 val_242 244 val_244 248 val_248 +252 val_252 +256 val_256 +256 val_256 +258 val_258 260 val_260 262 val_262 266 val_266 +272 val_272 +272 val_272 +274 val_274 +278 val_278 +278 val_278 280 val_280 280 val_280 282 val_282 @@ -2658,6 +2655,14 @@ 286 val_286 288 val_288 288 val_288 +292 val_292 +296 val_296 +298 val_298 +298 val_298 +298 val_298 +302 val_302 +306 val_306 +308 val_308 310 val_310 316 val_316 316 val_316 @@ -2665,12 +2670,34 @@ 318 val_318 318 val_318 318 val_318 +322 val_322 +322 val_322 332 val_332 336 val_336 338 val_338 +342 val_342 +342 val_342 +344 val_344 +344 val_344 +348 val_348 +348 val_348 +348 val_348 +348 val_348 +348 val_348 356 val_356 +360 val_360 +362 val_362 +364 val_364 +366 val_366 +368 val_368 374 val_374 378 val_378 +382 val_382 +382 val_382 +384 val_384 +384 val_384 +384 val_384 +386 val_386 392 val_392 394 val_394 396 val_396 @@ -2684,11 +2711,28 @@ 406 val_406 406 val_406 406 val_406 +414 val_414 +414 val_414 +418 val_418 424 val_424 424 val_424 +430 val_430 +430 val_430 +430 val_430 +432 val_432 +436 val_436 +438 val_438 +438 val_438 +438 val_438 444 val_444 446 val_446 448 val_448 +452 val_452 +454 val_454 +454 val_454 +454 val_454 +458 val_458 +458 val_458 460 val_460 462 val_462 462 val_462 @@ -2699,11 +2743,23 @@ 468 val_468 468 val_468 468 val_468 +470 val_470 +472 val_472 +478 val_478 +478 val_478 480 val_480 480 val_480 480 val_480 482 val_482 484 val_484 +490 val_490 +492 val_492 +492 val_492 +494 val_494 +496 val_496 +498 val_498 +498 val_498 +498 val_498 PREHOOK: query: EXPLAIN EXTENDED SELECT s.* FROM srcbucket2 TABLESAMPLE (BUCKET 2 OUT OF 4 on key) s ORDER BY key, value PREHOOK: type: QUERY @@ -2728,42 +2784,37 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 4) = 1) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 4) = 1) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2/srcbucket21.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2/srcbucket21.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 Partition - base file name: srcbucket21.txt + base file name: srcbucket2 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -2773,12 +2824,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 name srcbucket2 serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504428 + transient_lastDdlTime 1290519877 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -2790,12 +2841,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket2 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket2 name srcbucket2 serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504428 + transient_lastDdlTime 1290519877 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket2 name: srcbucket2 @@ -2804,9 +2855,9 @@ File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-25-16_389_3169479756049941851/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-38_178_1711030805393518113/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-25-16_389_3169479756049941851/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-38_178_1711030805393518113/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -2827,34 +2878,115 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket2 -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-16_781_3157391389694103452/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-38_315_5115954524988270253/-mr-10000 POSTHOOK: query: SELECT s.* FROM srcbucket2 TABLESAMPLE (BUCKET 2 OUT OF 4 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket2 -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-16_781_3157391389694103452/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-38_315_5115954524988270253/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 5 val_5 5 val_5 5 val_5 9 val_9 +17 val_17 +33 val_33 +37 val_37 +37 val_37 41 val_41 +53 val_53 +57 val_57 +65 val_65 +69 val_69 +77 val_77 85 val_85 +97 val_97 +97 val_97 +105 val_105 +113 val_113 +113 val_113 +125 val_125 +125 val_125 +129 val_129 +129 val_129 133 val_133 137 val_137 137 val_137 +145 val_145 +149 val_149 +149 val_149 +153 val_153 +157 val_157 +165 val_165 +165 val_165 +169 val_169 +169 val_169 +169 val_169 +169 val_169 177 val_177 +181 val_181 +189 val_189 +193 val_193 +193 val_193 +193 val_193 +197 val_197 +197 val_197 +201 val_201 +205 val_205 +205 val_205 +209 val_209 +209 val_209 +213 val_213 +213 val_213 +217 val_217 +217 val_217 221 val_221 221 val_221 229 val_229 229 val_229 +233 val_233 +233 val_233 +237 val_237 +237 val_237 +241 val_241 +249 val_249 +257 val_257 265 val_265 265 val_265 +273 val_273 +273 val_273 +273 val_273 +277 val_277 +277 val_277 +277 val_277 +277 val_277 +281 val_281 +281 val_281 +285 val_285 +289 val_289 +305 val_305 +309 val_309 +309 val_309 317 val_317 317 val_317 +321 val_321 +321 val_321 +325 val_325 +325 val_325 +333 val_333 +333 val_333 +341 val_341 +345 val_345 353 val_353 353 val_353 +365 val_365 +369 val_369 +369 val_369 +369 val_369 +373 val_373 +377 val_377 +389 val_389 393 val_393 397 val_397 397 val_397 @@ -2866,13 +2998,32 @@ 409 val_409 409 val_409 409 val_409 +413 val_413 +413 val_413 +417 val_417 +417 val_417 +417 val_417 +421 val_421 +429 val_429 +429 val_429 +437 val_437 449 val_449 +453 val_453 +457 val_457 +469 val_469 +469 val_469 +469 val_469 +469 val_469 +469 val_469 +477 val_477 481 val_481 485 val_485 489 val_489 489 val_489 489 val_489 489 val_489 +493 val_493 +497 val_497 PREHOOK: query: CREATE TABLE empty_bucket (key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE POSTHOOK: query: CREATE TABLE empty_bucket (key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE @@ -2904,44 +3055,80 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value + expr: _col1 type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string + sort order: ++ + tag: -1 + value expressions: + expr: _col0 + type: int + expr: _col1 + type: string Needs Tagging: false + Path -> Alias: + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/empty_bucket [s] + Path -> Partition: + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/empty_bucket + Partition + base file name: empty_bucket + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count 2 + bucket_field_name key + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/empty_bucket + name empty_bucket + serialization.ddl struct empty_bucket { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290519947 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count 2 + bucket_field_name key + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/empty_bucket + name empty_bucket + serialization.ddl struct empty_bucket { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290519947 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: empty_bucket + name: empty_bucket Reduce Operator Tree: Extract File Output Operator compressed: false GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-25-21_376_8476330244078104264/-ext-10001 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-47_726_6319788223237144597/-ext-10001 NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-25-21_376_8476330244078104264/-ext-10001/ + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-47_726_6319788223237144597/-ext-10001/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -2962,11 +3149,11 @@ ORDER BY key, value PREHOOK: type: QUERY PREHOOK: Input: default@empty_bucket -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-21_480_8277252386943120890/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-47_826_3824610300547913500/-mr-10000 POSTHOOK: query: SELECT s.* FROM empty_bucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) s ORDER BY key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@empty_bucket -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-21_480_8277252386943120890/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-45-47_826_3824610300547913500/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] Index: ql/src/test/results/clientpositive/sample7.q.out =================================================================== --- ql/src/test/results/clientpositive/sample7.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample7.q.out (working copy) @@ -37,56 +37,51 @@ Filter Operator isSamplingPred: false predicate: - expr: ((((hash(key) & 2147483647) % 4) = 0) and (key > 100)) + expr: (key > 100) type: boolean Filter Operator isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 4) = 0) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.types int:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 - name dest1 - serialization.ddl struct dest1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510325 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value + columns.types int:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 + name dest1 + serialization.ddl struct dest1 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1290520417 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt Partition base file name: srcbucket0.txt input format: org.apache.hadoop.mapred.TextInputFormat @@ -98,12 +93,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290520410 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -115,12 +110,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290520410 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -132,14 +127,14 @@ Move Operator files: hdfs directory: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10002 - destination: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10002 + destination: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10000 Stage: Stage-0 Move Operator tables: replace: true - source: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10000 + source: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -149,28 +144,28 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510325 + transient_lastDdlTime 1290520417 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 - tmp directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10001 + tmp directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10001 Stage: Stage-2 Stats-Aggr Operator - Stats Aggregation Key Prefix: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10000/ + Stats Aggregation Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10000/ Stage: Stage-3 Map Reduce Alias -> Map Operator Tree: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10002 File Output Operator compressed: false GlobalTableId: 0 - directory: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10000 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10000 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -181,12 +176,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510325 + transient_lastDdlTime 1290520417 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 TotalFiles: 1 @@ -194,9 +189,9 @@ MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10002 [pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10002] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10002 [pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10002] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/scratchdir/hive_2010-09-14_17-25-25_949_8398069099491911720/-ext-10002 + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_05-53-37_031_1506748942065140855/-ext-10002 Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat @@ -207,12 +202,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510325 + transient_lastDdlTime 1290520417 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -223,12 +218,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/dest1 + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dest1 name dest1 serialization.ddl struct dest1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284510325 + transient_lastDdlTime 1290520417 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dest1 name: dest1 @@ -251,11 +246,11 @@ PREHOOK: query: SELECT dest1.* FROM dest1 PREHOOK: type: QUERY PREHOOK: Input: default@dest1 -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-31_960_8165631533406459791/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-53-45_596_3764855055071931302/-mr-10000 POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-31_960_8165631533406459791/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-53-45_596_3764855055071931302/-mr-10000 POSTHOOK: Lineage: dest1.key SIMPLE [(srcbucket)s.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcbucket)s.FieldSchema(name:value, type:string, comment:null), ] 468 val_469 Index: ql/src/test/results/clientpositive/sample8.q.out =================================================================== --- ql/src/test/results/clientpositive/sample8.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample8.q.out (working copy) @@ -33,49 +33,39 @@ alias: s GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 1) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 1) = 0) - type: boolean - Reduce Output Operator - sort order: - tag: 0 - value expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string + Reduce Output Operator + sort order: + tag: 0 + value expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string t TableScan alias: t GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 10) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 10) = 0) - type: boolean - Reduce Output Operator - sort order: - tag: 1 - value expressions: - expr: key - type: string - expr: value - type: string + Reduce Output Operator + sort order: + tag: 1 + value expressions: + expr: key + type: string + expr: value + type: string Needs Tagging: true Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [t, s] Index: ql/src/test/results/clientpositive/sample9.q.out =================================================================== --- ql/src/test/results/clientpositive/sample9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/sample9.q.out (working copy) @@ -22,52 +22,47 @@ alias: a GatherStats: false Filter Operator - isSamplingPred: false + isSamplingPred: true predicate: expr: (((hash(key) & 2147483647) % 2) = 0) type: boolean - Filter Operator - isSamplingPred: true - predicate: - expr: (((hash(key) & 2147483647) % 2) = 0) - type: boolean + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: key + expr: _col0 type: int - expr: value + expr: _col1 type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-14_17-25-46_949_6966788619794916428/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-14_17-25-46_949_6966788619794916428/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1 - columns.types int:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-47-44_363_7942337067051626222/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-47-44_363_7942337067051626222/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1 + columns.types int:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Needs Tagging: false Path -> Alias: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt [s:a] + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket [s:a] Path -> Partition: - pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket Partition - base file name: srcbucket0.txt + base file name: srcbucket input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -77,12 +72,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290520057 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -94,12 +89,12 @@ columns.types int:string file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcbucket + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/srcbucket name srcbucket serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1284504426 + transient_lastDdlTime 1290520057 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: srcbucket name: srcbucket @@ -113,12 +108,12 @@ FROM (SELECT a.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) a) s PREHOOK: type: QUERY PREHOOK: Input: default@srcbucket -PREHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-47_211_4954933153324523917/-mr-10000 +PREHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-47-44_749_661756637908541876/-mr-10000 POSTHOOK: query: SELECT s.* FROM (SELECT a.* FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 2 on key) a) s POSTHOOK: type: QUERY POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: file:/tmp/nzhang/hive_2010-09-14_17-25-47_211_4954933153324523917/-mr-10000 +POSTHOOK: Output: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-23_05-47-44_749_661756637908541876/-mr-10000 474 val_475 62 val_63 468 val_469 Index: ql/src/test/results/clientpositive/semijoin.q.out =================================================================== --- ql/src/test/results/clientpositive/semijoin.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/semijoin.q.out (working copy) @@ -517,33 +517,29 @@ predicate: expr: (key < '15') type: boolean - Filter Operator - predicate: - expr: (key < '15') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: int + outputColumnNames: key + Group By Operator + bucketGroup: false + keys: expr: key type: int - outputColumnNames: key - Group By Operator - bucketGroup: false - keys: - expr: key + expr: key + type: int + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col1 type: int - expr: key + sort order: + + Map-reduce partition columns: + expr: _col1 type: int - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col1 - type: int - sort order: + - Map-reduce partition columns: - expr: _col1 - type: int - tag: 1 + tag: 1 Reduce Operator Tree: Join Operator condition map: @@ -653,35 +649,31 @@ predicate: expr: (value < 'val_10') type: boolean - Filter Operator - predicate: - expr: (value < 'val_10') - type: boolean - Select Operator - expressions: + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: key, value + Group By Operator + bucketGroup: false + keys: expr: key type: int expr: value type: string - outputColumnNames: key, value - Group By Operator - bucketGroup: false - keys: - expr: key + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: + expr: _col0 type: int - expr: value - type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 1 + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: 1 Reduce Operator Tree: Join Operator condition map: @@ -787,36 +779,32 @@ predicate: expr: (key > 5) type: boolean - Filter Operator - predicate: - expr: (key > 5) - type: boolean + Select Operator + expressions: + expr: key + type: int + outputColumnNames: _col0 Select Operator expressions: - expr: key + expr: _col0 type: int outputColumnNames: _col0 - Select Operator - expressions: + Group By Operator + bucketGroup: false + keys: expr: _col0 type: int + mode: hash outputColumnNames: _col0 - Group By Operator - bucketGroup: false - keys: + Reduce Output Operator + key expressions: expr: _col0 type: int - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 1 + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: 1 Reduce Operator Tree: Join Operator condition map: @@ -916,46 +904,38 @@ predicate: expr: ((key > 5) and (value <= 'val_20')) type: boolean - Filter Operator - predicate: - expr: (key > 5) - type: boolean + Select Operator + expressions: + expr: key + type: int + expr: value + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: key + expr: _col0 type: int - expr: value + expr: _col1 type: string outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col1 <= 'val_20') - type: boolean - Select Operator - expressions: + Group By Operator + bucketGroup: false + keys: + expr: _col0 + type: int + expr: _col1 + type: string + mode: hash + outputColumnNames: _col0, _col1 + Reduce Output Operator + key expressions: expr: _col0 type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - Group By Operator - bucketGroup: false - keys: - expr: _col0 - type: int - expr: _col1 - type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 1 + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: 1 Reduce Operator Tree: Join Operator condition map: @@ -1054,36 +1034,32 @@ predicate: expr: (key > 2) type: boolean - Filter Operator - predicate: - expr: (key > 2) - type: boolean + Select Operator + expressions: + expr: key + type: int + outputColumnNames: _col0 Select Operator expressions: - expr: key + expr: _col0 type: int outputColumnNames: _col0 - Select Operator - expressions: + Group By Operator + bucketGroup: false + keys: expr: _col0 type: int + mode: hash outputColumnNames: _col0 - Group By Operator - bucketGroup: false - keys: + Reduce Output Operator + key expressions: expr: _col0 type: int - mode: hash - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 1 + sort order: + + Map-reduce partition columns: + expr: _col0 + type: int + tag: 1 Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/skewjoin.q.out =================================================================== --- ql/src/test/results/clientpositive/skewjoin.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/skewjoin.q.out (working copy) @@ -1194,22 +1194,18 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 100) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 0 - value expressions: - expr: _col0 - type: string + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 0 + value expressions: + expr: _col0 + type: string src2:src TableScan alias: src @@ -1244,19 +1240,15 @@ expr: key type: string outputColumnNames: _col0 - Filter Operator - predicate: - expr: (_col0 < 80) - type: boolean - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: 2 + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + Map-reduce partition columns: + expr: _col0 + type: string + tag: 2 Reduce Operator Tree: Join Operator condition map: Index: ql/src/test/results/clientpositive/smb_mapjoin9.q.out =================================================================== --- ql/src/test/results/clientpositive/smb_mapjoin9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/smb_mapjoin9.q.out (working copy) @@ -88,27 +88,23 @@ expr: _col7 type: string outputColumnNames: _col0, _col2, _col5, _col6, _col7 - Filter Operator - predicate: - expr: (((_col2 = '2010-10-15') and (_col7 = '2010-10-15')) and _col5 is not null) - type: boolean - Select Operator - expressions: - expr: _col5 - type: int - expr: _col6 - type: string - expr: _col7 - type: string - expr: _col0 - type: int - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col5 + type: int + expr: _col6 + type: string + expr: _col7 + type: string + expr: _col0 + type: int + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/smb_mapjoin_6.q.out =================================================================== --- ql/src/test/results/clientpositive/smb_mapjoin_6.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/smb_mapjoin_6.q.out (working copy) @@ -2634,29 +2634,25 @@ expr: _col5 type: string outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: (_col0 > 1000) - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col4 - type: int - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: smb_join_results + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col4 + type: int + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: smb_join_results Stage: Stage-0 Move Operator @@ -2790,29 +2786,25 @@ expr: _col5 type: string outputColumnNames: _col0, _col1, _col4, _col5 - Filter Operator - predicate: - expr: (_col0 > 1000) - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col4 - type: int - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: smb_join_results + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col4 + type: int + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: smb_join_results Stage: Stage-0 Move Operator @@ -2958,31 +2950,27 @@ expr: _col9 type: string outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9 - Filter Operator - predicate: - expr: (_col0 > 1000) - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col4 - type: int - expr: _col5 - type: string - expr: _col8 - type: int - expr: _col9 - type: string - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col4 + type: int + expr: _col5 + type: string + expr: _col8 + type: int + expr: _col9 + type: string + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/stats11.q.out =================================================================== --- ql/src/test/results/clientpositive/stats11.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/stats11.q.out (working copy) @@ -152,46 +152,41 @@ expr: _col6 type: string outputColumnNames: _col0, _col1, _col5, _col6 - Filter Operator - isSamplingPred: false - predicate: - expr: (_col6 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_20-31-38_084_5882852540626894180/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_20-31-38_084_5882852540626894180/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - transient_lastDdlTime 1289277098 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_02-02-42_241_2760045756897471406/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_02-02-42_241_2760045756897471406/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + transient_lastDdlTime 1289988162 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false @@ -583,50 +578,45 @@ expr: _col6 type: string outputColumnNames: _col0, _col1, _col5, _col6 - Filter Operator - isSamplingPred: false - predicate: - expr: (_col6 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col5 - type: string - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_20-32-35_603_7737558187075598476/-ext-10002 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_20-32-35_603_7737558187075598476/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value1,value2 - columns.types string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/bucketmapjoin_tmp_result - name bucketmapjoin_tmp_result - numFiles 1 - numPartitions 0 - numRows 464 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 - transient_lastDdlTime 1289277139 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bucketmapjoin_tmp_result - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Select Operator + expressions: + expr: _col0 + type: int + expr: _col1 + type: string + expr: _col5 + type: string + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_02-04-01_017_3109478588671681748/-ext-10002 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-17_02-04-01_017_3109478588671681748/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns key,value1,value2 + columns.types string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/bucketmapjoin_tmp_result + name bucketmapjoin_tmp_result + numFiles 1 + numPartitions 0 + numRows 464 + serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 8983 + transient_lastDdlTime 1289988216 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: bucketmapjoin_tmp_result + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Local Work: Map Reduce Local Work Needs Tagging: false Index: ql/src/test/results/clientpositive/stats2.q.out =================================================================== --- ql/src/test/results/clientpositive/stats2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/stats2.q.out (working copy) @@ -23,29 +23,25 @@ srcpart TableScan alias: srcpart - Filter Operator - predicate: - expr: ds is not null - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: analyze_t1 + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: analyze_t1 Stage: Stage-0 Move Operator Index: ql/src/test/results/clientpositive/subq.q.out =================================================================== --- ql/src/test/results/clientpositive/subq.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/subq.q.out (working copy) @@ -31,30 +31,26 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 Select Operator expressions: - expr: key + expr: _col0 type: string - expr: value + expr: _col1 type: string outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-4 Conditional Operator Index: ql/src/test/results/clientpositive/subq2.q.out =================================================================== --- ql/src/test/results/clientpositive/subq2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/subq2.q.out (working copy) @@ -69,23 +69,19 @@ expr: _col1 type: bigint outputColumnNames: _col0, _col1 - Filter Operator - predicate: - expr: (_col0 >= 90) - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: bigint - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: bigint + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/transform_ppr1.q.out =================================================================== --- ql/src/test/results/clientpositive/transform_ppr1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/transform_ppr1.q.out (working copy) @@ -50,22 +50,27 @@ columns.types string,string,string field.delim 9 serialization.format 9 - Reduce Output Operator - key expressions: - expr: _col1 - type: string - sort order: + - Map-reduce partition columns: - expr: _col1 - type: string - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string + Filter Operator + isSamplingPred: false + predicate: + expr: ((_col1 < 100) and (_col0 = '2008-04-08')) + type: boolean + Reduce Output Operator + key expressions: + expr: _col1 + type: string + sort order: + + Map-reduce partition columns: + expr: _col1 + type: string + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/nzhang/work/784/apache-hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [tmap:src] @@ -239,34 +244,29 @@ name: srcpart Reduce Operator Tree: Extract - Filter Operator - isSamplingPred: false - predicate: - expr: ((_col1 < 100) and (_col0 = '2008-04-08')) - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/nzhang/hive_2010-09-15_17-09-49_113_6975059842786318267/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/nzhang/hive_2010-09-15_17-09-49_113_6975059842786318267/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1 - columns.types string:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_02-12-56_302_8393123164421135469/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_02-12-56_302_8393123164421135469/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1 + columns.types string:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/transform_ppr2.q.out =================================================================== --- ql/src/test/results/clientpositive/transform_ppr2.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/transform_ppr2.q.out (working copy) @@ -33,30 +33,30 @@ TableScan alias: src GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (ds = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: ds - type: string - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1, _col2 - Transform Operator - command: /bin/cat - output info: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1,_col2 - columns.types string,string,string - field.delim 9 - serialization.format 9 + Select Operator + expressions: + expr: ds + type: string + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1, _col2 + Transform Operator + command: /bin/cat + output info: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1,_col2 + columns.types string,string,string + field.delim 9 + serialization.format 9 + Filter Operator + isSamplingPred: false + predicate: + expr: (_col1 < 100) + type: boolean Reduce Output Operator key expressions: expr: _col1 @@ -162,34 +162,29 @@ name: srcpart Reduce Operator Tree: Extract - Filter Operator - isSamplingPred: false - predicate: - expr: (_col1 < 100) - type: boolean - Select Operator - expressions: - expr: _col1 - type: string - expr: _col2 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - directory: file:/tmp/sdong/hive_2010-10-29_16-49-00_753_64980182722766710/-ext-10001 - NumFilesPerFileSink: 1 - Stats Publishing Key Prefix: file:/tmp/sdong/hive_2010-10-29_16-49-00_753_64980182722766710/-ext-10001/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns _col0,_col1 - columns.types string:string - serialization.format 1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + Select Operator + expressions: + expr: _col1 + type: string + expr: _col2 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + directory: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_02-13-06_009_2732362870185708504/-ext-10001 + NumFilesPerFileSink: 1 + Stats Publishing Key Prefix: file:/var/folders/uc/ucuNeMAVGQGzy3459D8z2+++Z0Q/-Tmp-/amarsri/hive_2010-11-17_02-13-06_009_2732362870185708504/-ext-10001/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + columns _col0,_col1 + columns.types string:string + serialization.format 1 + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf1.q.out =================================================================== --- ql/src/test/results/clientpositive/udf1.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf1.q.out (working copy) @@ -51,61 +51,57 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: ('a' like '%a%') - type: boolean - expr: ('b' like '%a%') - type: boolean - expr: ('ab' like '%a%') - type: boolean - expr: ('ab' like '%a_') - type: boolean - expr: ('%_' like '\%\_') - type: boolean - expr: ('ab' like '\%\_') - type: boolean - expr: ('ab' like '_a%') - type: boolean - expr: ('ab' like 'a') - type: boolean - expr: ('' rlike '.*') - type: boolean - expr: ('a' rlike '[ab]') - type: boolean - expr: ('' rlike '[ab]') - type: boolean - expr: ('hadoop' rlike '[a-z]*') - type: boolean - expr: ('hadoop' rlike 'o*') - type: boolean - expr: regexp_replace('abc', 'b', 'c') - type: string - expr: regexp_replace('abc', 'z', 'a') - type: string - expr: regexp_replace('abbbb', 'bb', 'b') - type: string - expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive') - type: string - expr: regexp_replace('hadoopAAA', 'A.*', '') - type: string - expr: regexp_replace('abc', '', 'A') - type: string - expr: ('abc' rlike '') - type: boolean - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + Select Operator + expressions: + expr: ('a' like '%a%') + type: boolean + expr: ('b' like '%a%') + type: boolean + expr: ('ab' like '%a%') + type: boolean + expr: ('ab' like '%a_') + type: boolean + expr: ('%_' like '\%\_') + type: boolean + expr: ('ab' like '\%\_') + type: boolean + expr: ('ab' like '_a%') + type: boolean + expr: ('ab' like 'a') + type: boolean + expr: ('' rlike '.*') + type: boolean + expr: ('a' rlike '[ab]') + type: boolean + expr: ('' rlike '[ab]') + type: boolean + expr: ('hadoop' rlike '[a-z]*') + type: boolean + expr: ('hadoop' rlike 'o*') + type: boolean + expr: regexp_replace('abc', 'b', 'c') + type: string + expr: regexp_replace('abc', 'z', 'a') + type: string + expr: regexp_replace('abbbb', 'bb', 'b') + type: string + expr: regexp_replace('hadoop', '(.)[a-z]*', '$1ive') + type: string + expr: regexp_replace('hadoopAAA', 'A.*', '') + type: string + expr: regexp_replace('abc', '', 'A') + type: string + expr: ('abc' rlike '') + type: boolean + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/udf9.q.out =================================================================== --- ql/src/test/results/clientpositive/udf9.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf9.q.out (working copy) @@ -38,51 +38,47 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: datediff('2008-12-31', '2009-01-01') - type: int - expr: datediff('2008-03-01', '2008-02-28') - type: int - expr: datediff('2007-03-01', '2007-01-28') - type: int - expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00') - type: int - expr: date_add('2008-12-31', 1) - type: string - expr: date_add('2008-12-31', 365) - type: string - expr: date_add('2008-02-28', 2) - type: string - expr: date_add('2009-02-28', 2) - type: string - expr: date_add('2007-02-28', 365) - type: string - expr: date_add('2007-02-28 23:59:59', 730) - type: string - expr: date_sub('2009-01-01', 1) - type: string - expr: date_sub('2009-01-01', 365) - type: string - expr: date_sub('2008-02-28', 2) - type: string - expr: date_sub('2009-02-28', 2) - type: string - expr: date_sub('2007-02-28', 365) - type: string - expr: date_sub('2007-02-28 01:12:34', 730) - type: string - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: datediff('2008-12-31', '2009-01-01') + type: int + expr: datediff('2008-03-01', '2008-02-28') + type: int + expr: datediff('2007-03-01', '2007-01-28') + type: int + expr: datediff('2008-03-01 23:59:59', '2008-03-02 00:00:00') + type: int + expr: date_add('2008-12-31', 1) + type: string + expr: date_add('2008-12-31', 365) + type: string + expr: date_add('2008-02-28', 2) + type: string + expr: date_add('2009-02-28', 2) + type: string + expr: date_add('2007-02-28', 365) + type: string + expr: date_add('2007-02-28 23:59:59', 730) + type: string + expr: date_sub('2009-01-01', 1) + type: string + expr: date_sub('2009-01-01', 365) + type: string + expr: date_sub('2008-02-28', 2) + type: string + expr: date_sub('2009-02-28', 2) + type: string + expr: date_sub('2007-02-28', 365) + type: string + expr: date_sub('2007-02-28 01:12:34', 730) + type: string + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_10_trims.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_10_trims.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_10_trims.q.out (working copy) @@ -37,23 +37,19 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim(' abc ')))))))))) - type: string - outputColumnNames: _col0 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dest1 + Select Operator + expressions: + expr: trim(trim(trim(trim(trim(trim(trim(trim(trim(trim(' abc ')))))))))) + type: string + outputColumnNames: _col0 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dest1 Stage: Stage-5 Conditional Operator Index: ql/src/test/results/clientpositive/udf_hour.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_hour.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_hour.q.out (working copy) @@ -40,25 +40,21 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: hour('2009-08-07 13:14:15') - type: int - expr: hour('13:14:15') - type: int - expr: hour('2009-08-07') - type: int - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: hour('2009-08-07 13:14:15') + type: int + expr: hour('13:14:15') + type: int + expr: hour('2009-08-07') + type: int + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_isnull_isnotnull.q.out (working copy) @@ -50,26 +50,22 @@ predicate: expr: true is not null type: boolean - Filter Operator - predicate: - expr: true is not null - type: boolean - Select Operator - expressions: - expr: null is null - type: boolean - expr: 1 is not null - type: boolean - expr: 'my string' is not null - type: boolean - outputColumnNames: _col0, _col1, _col2 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: null is null + type: boolean + expr: 1 is not null + type: boolean + expr: 'my string' is not null + type: boolean + outputColumnNames: _col0, _col1, _col2 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -127,26 +123,22 @@ predicate: expr: (lint is not null and (not mstringstring is null)) type: boolean - Filter Operator - predicate: - expr: (lint is not null and (not mstringstring is null)) - type: boolean - Select Operator - expressions: - expr: lint is not null - type: boolean - expr: lintstring is not null - type: boolean - expr: mstringstring is not null - type: boolean - outputColumnNames: _col0, _col1, _col2 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: lint is not null + type: boolean + expr: lintstring is not null + type: boolean + expr: mstringstring is not null + type: boolean + outputColumnNames: _col0, _col1, _col2 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_like.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_like.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_like.q.out (working copy) @@ -41,43 +41,39 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: ('_%_' like '%\_\%\_%') - type: boolean - expr: ('__' like '%\_\%\_%') - type: boolean - expr: ('%%_%_' like '%\_\%\_%') - type: boolean - expr: ('%_%_%' like '%\%\_\%') - type: boolean - expr: ('_%_' like '\%\_%') - type: boolean - expr: ('%__' like '__\%%') - type: boolean - expr: ('_%' like '\_\%\_\%%') - type: boolean - expr: ('_%' like '\_\%_%') - type: boolean - expr: ('%_' like '\%\_') - type: boolean - expr: ('ab' like '\%\_') - type: boolean - expr: ('ab' like '_a%') - type: boolean - expr: ('ab' like 'a') - type: boolean - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: ('_%_' like '%\_\%\_%') + type: boolean + expr: ('__' like '%\_\%\_%') + type: boolean + expr: ('%%_%_' like '%\_\%\_%') + type: boolean + expr: ('%_%_%' like '%\%\_\%') + type: boolean + expr: ('_%_' like '\%\_%') + type: boolean + expr: ('%__' like '__\%%') + type: boolean + expr: ('_%' like '\_\%\_\%%') + type: boolean + expr: ('_%' like '\_\%_%') + type: boolean + expr: ('%_' like '\%\_') + type: boolean + expr: ('ab' like '\%\_') + type: boolean + expr: ('ab' like '_a%') + type: boolean + expr: ('ab' like 'a') + type: boolean + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_lower.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_lower.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_lower.q.out (working copy) @@ -36,23 +36,19 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: lower('AbC 123') - type: string - expr: upper('AbC 123') - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: lower('AbC 123') + type: string + expr: upper('AbC 123') + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_minute.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_minute.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_minute.q.out (working copy) @@ -40,25 +40,21 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: minute('2009-08-07 13:14:15') - type: int - expr: minute('13:14:15') - type: int - expr: minute('2009-08-07') - type: int - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: minute('2009-08-07 13:14:15') + type: int + expr: minute('13:14:15') + type: int + expr: minute('2009-08-07') + type: int + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_notequal.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_notequal.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_notequal.q.out (working copy) @@ -48,23 +48,19 @@ predicate: expr: (key <> '302') type: boolean - Filter Operator - predicate: - expr: (key <> '302') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator @@ -610,23 +606,19 @@ predicate: expr: (key <> '302') type: boolean - Filter Operator - predicate: - expr: (key <> '302') - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_parse_url.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_parse_url.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_parse_url.q.out (working copy) @@ -63,41 +63,37 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'HOST') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PATH') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'REF') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k2') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k1') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k3') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'FILE') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PROTOCOL') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'USERINFO') - type: string - expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY') - type: string - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'HOST') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PATH') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'REF') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k2') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k1') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'QUERY', 'k3') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'FILE') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'PROTOCOL') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'USERINFO') + type: string + expr: parse_url('http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1', 'AUTHORITY') + type: string + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_second.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_second.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_second.q.out (working copy) @@ -40,25 +40,21 @@ predicate: expr: (key = 86) type: boolean - Filter Operator - predicate: - expr: (key = 86) - type: boolean - Select Operator - expressions: - expr: second('2009-08-07 13:14:15') - type: int - expr: second('13:14:15') - type: int - expr: second('2009-08-07') - type: int - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: second('2009-08-07 13:14:15') + type: int + expr: second('13:14:15') + type: int + expr: second('2009-08-07') + type: int + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/udf_size.q.out =================================================================== --- ql/src/test/results/clientpositive/udf_size.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/udf_size.q.out (working copy) @@ -44,28 +44,24 @@ predicate: expr: (lint is not null and (not mstringstring is null)) type: boolean - Filter Operator - predicate: - expr: (lint is not null and (not mstringstring is null)) - type: boolean - Select Operator - expressions: - expr: size(lint) - type: int - expr: size(lintstring) - type: int - expr: size(mstringstring) - type: int - expr: size(null) - type: int - outputColumnNames: _col0, _col1, _col2, _col3 - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: size(lint) + type: int + expr: size(lintstring) + type: int + expr: size(mstringstring) + type: int + expr: size(null) + type: int + outputColumnNames: _col0, _col1, _col2, _col3 + Limit + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-0 Fetch Operator Index: ql/src/test/results/clientpositive/union.q.out =================================================================== --- ql/src/test/results/clientpositive/union.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/union.q.out (working copy) @@ -39,31 +39,27 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat null-subquery2:unioninput-subquery2:src TableScan alias: src @@ -71,31 +67,27 @@ predicate: expr: (key > 100) type: boolean - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Union - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 1 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Stage: Stage-4 Conditional Operator Index: ql/src/test/results/clientpositive/union20.q.out =================================================================== --- ql/src/test/results/clientpositive/union20.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/union20.q.out (working copy) @@ -164,23 +164,19 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-4 Map Reduce @@ -192,23 +188,19 @@ predicate: expr: (key < 10) type: boolean - Filter Operator - predicate: - expr: (key < 10) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + outputColumnNames: _col0, _col1 + File Output Operator + compressed: false + GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat Stage: Stage-5 Map Reduce Index: ql/src/test/results/clientpositive/union22.q.out =================================================================== --- ql/src/test/results/clientpositive/union22.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/union22.q.out (working copy) @@ -105,32 +105,27 @@ predicate: expr: (k0 > 50) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '1') and (k0 > 50)) - type: boolean - Select Operator - expressions: - expr: k1 - type: string - expr: k3 - type: string - expr: k4 - type: string - outputColumnNames: _col1, _col3, _col4 - HashTable Sink Operator - condition expressions: - 0 {k1} {k2} - 1 {_col3} {_col4} - filter predicates: - 0 {(ds = '1')} - 1 - handleSkewJoin: false - keys: - 0 [Column[k1]] - 1 [Column[_col1]] - Position of Big Table: 0 + Select Operator + expressions: + expr: k1 + type: string + expr: k3 + type: string + expr: k4 + type: string + outputColumnNames: _col1, _col3, _col4 + HashTable Sink Operator + condition expressions: + 0 {k1} {k2} + 1 {_col3} {_col4} + filter predicates: + 0 {(ds = '1')} + 1 + handleSkewJoin: false + keys: + 0 [Column[k1]] + 1 [Column[_col1]] + Position of Big Table: 0 Stage: Stage-1 Map Reduce @@ -244,20 +239,79 @@ expr: _col11 type: string outputColumnNames: _col0, _col1, _col10, _col11 + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col10 + type: string + expr: _col11 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + GlobalTableId: 1 + directory: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_06-23-20_540_8033306049764991055/-ext-10000 + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2/ + Stats Publishing Key Prefix: pfile:/Users/amarsri/Documents/workspace/hive/build/ql/scratchdir/hive_2010-11-23_06-23-20_540_8033306049764991055/-ext-10000/ + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + columns k1,k2,k3,k4 + columns.types string:string:string:string + file.inputformat org.apache.hadoop.mapred.TextInputFormat + file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + location pfile:/Users/amarsri/Documents/workspace/hive/build/ql/test/data/warehouse/dst_union22 + name dst_union22 + numFiles 1 + numPartitions 1 + numRows 500 + partition_columns ds + serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 11624 + transient_lastDdlTime 1290522190 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: dst_union22 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + null-subquery1:subq-subquery1:dst_union22_delta + TableScan + alias: dst_union22_delta + GatherStats: false Filter Operator isSamplingPred: false predicate: - expr: (_col0 > 20) + expr: (k0 <= 50) type: boolean Select Operator expressions: - expr: _col0 + expr: k1 type: string - expr: _col1 + expr: k2 type: string - expr: _col10 + expr: k3 type: string - expr: _col11 + expr: k4 type: string outputColumnNames: _col0, _col1, _col2, _col3 Union @@ -304,75 +358,6 @@ TotalFiles: 1 GatherStats: true MultiFileSpray: false - null-subquery1:subq-subquery1:dst_union22_delta - TableScan - alias: dst_union22_delta - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: - expr: (k0 <= 50) - type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: ((ds = '1') and (k0 <= 50)) - type: boolean - Select Operator - expressions: - expr: k1 - type: string - expr: k2 - type: string - expr: k3 - type: string - expr: k4 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Union - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - GlobalTableId: 1 - directory: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_21-11-17_290_439394066953047897/-ext-10000 - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2/ - Stats Publishing Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_21-11-17_290_439394066953047897/-ext-10000/ - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns k1,k2,k3,k4 - columns.types string:string:string:string - file.inputformat org.apache.hadoop.mapred.TextInputFormat - file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - location pfile:/data/users/liyintang/hive-1754/build/ql/test/data/warehouse/dst_union22 - name dst_union22 - numFiles 1 - numPartitions 1 - numRows 500 - partition_columns ds - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 - transient_lastDdlTime 1289279470 - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: dst_union22 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false Needs Tagging: false Path -> Alias: file:/tmp/liyintang/hive_2010-11-08_21-11-17_290_439394066953047897/-mr-10002 [file:/tmp/liyintang/hive_2010-11-08_21-11-17_290_439394066953047897/-mr-10002] @@ -478,6 +463,7 @@ Stats-Aggr Operator Stats Aggregation Key Prefix: pfile:/data/users/liyintang/hive-1754/build/ql/scratchdir/hive_2010-11-08_21-11-17_290_439394066953047897/-ext-10000/ + PREHOOK: query: insert overwrite table dst_union22 partition (ds='2') select * from ( Index: ql/src/test/results/clientpositive/union_ppr.q.out =================================================================== --- ql/src/test/results/clientpositive/union_ppr.q.out (revision 1038445) +++ ql/src/test/results/clientpositive/union_ppr.q.out (working copy) @@ -36,54 +36,44 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Union - Filter Operator - isSamplingPred: false - predicate: - expr: (_col2 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string null-subquery2:a-subquery2:y TableScan alias: y @@ -93,54 +83,44 @@ predicate: expr: (key < 100) type: boolean - Filter Operator - isSamplingPred: false - predicate: - expr: (key < 100) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - expr: ds - type: string - expr: hr - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Union - Filter Operator - isSamplingPred: false - predicate: - expr: (_col2 = '2008-04-08') - type: boolean - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1, _col2, _col3 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - expr: _col2 - type: string - expr: _col3 - type: string + Select Operator + expressions: + expr: key + type: string + expr: value + type: string + expr: ds + type: string + expr: hr + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Union + Select Operator + expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string + outputColumnNames: _col0, _col1, _col2, _col3 + Reduce Output Operator + key expressions: + expr: _col0 + type: string + sort order: + + tag: -1 + value expressions: + expr: _col0 + type: string + expr: _col1 + type: string + expr: _col2 + type: string + expr: _col3 + type: string Needs Tagging: false Path -> Alias: pfile:/data/users/sdong/www/hive-trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [null-subquery1:a-subquery1:x, null-subquery2:a-subquery2:y] Index: ql/src/test/results/compiler/plan/case_sensitivity.q.xml =================================================================== --- ql/src/test/results/compiler/plan/case_sensitivity.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/case_sensitivity.q.xml (working copy) @@ -570,327 +570,176 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-31-00_262_886649804432989556/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-31-00_262_886649804432989556/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - - - - - - - lintstring - - - src_thrift - - - - - - - - - myint - - - mystring - - - underscore_int - - - - - - - - - int - - - - - - - - - - - - - - - - - - - - - - - - 0 - - - - - - - - - - - - - - - MYSTRING - - - false - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-51-35_606_6683476852621996377/-ext-10002 - - _col0 - - - - - - - lint - - - src_thrift - - - - - - - - - - - - - - - - - 1 - - - - - - - - - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-51-35_606_6683476852621996377/-ext-10000/ + + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - + + + + + + + + + + _col1 + + + + - - - _col0 + + + lintstring - - + + src_thrift + + + + + + + + myint + + + mystring + + + underscore_int + + + + + + + + + int + + + + + + + + + + + + + + + - - - _col1 + + + - - + + 0 + + + + + + + + MYSTRING + + + false + + + + - - - - - - + + _col0 + - - - - - - - lint - - - src_thrift - - - - - - - - - - - - - 0 - - - - + + + lint - - + + src_thrift - + + + + + @@ -900,44 +749,64 @@ - 0 + 1 - + - - - boolean - - + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -951,28 +820,22 @@ - + - lint + _col0 - - src_thrift - - + - + - lintstring + _col1 - - src_thrift - - + @@ -1043,7 +906,11 @@ - + + + boolean + + @@ -1106,7 +973,17 @@ - + + + lint + + + src_thrift + + + + + @@ -1126,7 +1003,17 @@ - + + + lintstring + + + src_thrift + + + + + Index: ql/src/test/results/compiler/plan/cast1.q.xml =================================================================== --- ql/src/test/results/compiler/plan/cast1.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/cast1.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011470 + 1290531098 @@ -107,592 +107,79 @@ - + - - - - - - - - - file:/tmp/njain/hive_2010-11-17_08-31-13_121_5524222101400973460/-ext-10001 - - - 1 - - - file:/tmp/njain/hive_2010-11-17_08-31-13_121_5524222101400973460/-ext-10001/ - - - - - org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - org.apache.hadoop.mapred.TextInputFormat - - - org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - - columns - _col0,_col1,_col2,_col3,_col4,_col5,_col6 - - - serialization.format - 1 - - - columns.types - int:double:double:double:int:boolean:int - - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - - - - - _col0 - - - - - - - - int - - - - - - - - - _col1 - - - - - - - - double - - - - - - - - - _col2 - - - - - - - - - - - - - _col3 - - - - - - - - - - - - - _col4 - - - - - - - - - - - - - _col5 - - - - - - - - boolean - - - - - - - - - _col6 - - - - - - - - - - - - - - + + + + + file:/tmp/amarsri/hive_2010-11-23_08-51-39_731_7334373255903755310/-ext-10001 - - - - - - _col6 - - - - - - - - - - true - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFToInteger - - - UDFToInteger - - - - - - - + + 1 - - _col5 - - - - - - - - - - 1 - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFToBoolean - - - UDFToBoolean - - - - - - - + + file:/tmp/amarsri/hive_2010-11-23_08-51-39_731_7334373255903755310/-ext-10001/ - - _col4 - - - - - - - - - - 3 - - - - - - - - - - - - - - 2.0 - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFToInteger - - - UDFToInteger - - - - - - - - - + + + + org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPPlus - - - + - - + + org.apache.hadoop.mapred.TextInputFormat - - + + org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - _col3 - - - - - - - - - - 3.0 - - + + + + columns + _col0,_col1,_col2,_col3,_col4,_col5,_col6 - - - - - - - 2.0 - - + + serialization.format + 1 - - - - - - true + + columns.types + int:double:double:double:int:boolean:int - - org.apache.hadoop.hive.ql.udf.UDFOPPlus - - - + - - - - - - _col2 - - - - - - - - - - 3 - - - - - - - - - - 2.0 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPPlus - - - + - - - - - - - + + 1 - - _col1 - - - - - - - - - - 3.0 - - - - - - - - - - 2 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPPlus - - - + - - - - - - - - - - _col0 - - - - - - - - - - 3 - - - - - - - - - - 2 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPPlus - - - + - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - _col1 - - - _col2 - - - _col3 - - - _col4 - - - _col5 - - - _col6 - - - - - - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + @@ -702,38 +189,46 @@ - - _c0 - _col0 + + + - + + + int + + - - _c1 - _col1 + + + - + + + double + + - - _c2 - _col2 + + + @@ -741,12 +236,12 @@ - - _c3 - _col3 + + + @@ -754,12 +249,12 @@ - - _c4 - _col4 + + + @@ -767,25 +262,29 @@ - - _c5 - _col5 + + + - + + + boolean + + - - _c6 - _col6 + + + @@ -799,69 +298,392 @@ - - - - + + + + _col6 + - - - key + + + - - src + + true + + + + + + + + org.apache.hadoop.hive.ql.udf.UDFToInteger + + + UDFToInteger + + + + + + + + + + _col5 + + + + + - - - string + + + + 1 + + + + + + + + + org.apache.hadoop.hive.ql.udf.UDFToBoolean + + + UDFToBoolean + + + + + + + + + + _col4 + + + + + + + + + + 3 + + + + + + + + + + + + + + 2.0 + + + + + + org.apache.hadoop.hive.ql.udf.UDFToInteger + + + UDFToInteger + + + + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFOPPlus + + + + + + + + + + + + + + _col3 + + + + + + + 3.0 + + + + + + + + + + 2.0 + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFOPPlus + + + + + + + + + + + + + + _col2 + + + + + + - 86 + 3 + + + + + + + 2.0 + + + - + + + true + + + org.apache.hadoop.hive.ql.udf.UDFOPPlus + + + + + + - + + + _col1 + + + + + + + + + + 3.0 + + + + + + + + + + 2 + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFOPPlus + + + + + + + + + + + + + + _col0 + + + + + + + + + + 3 + + + + + + + + + + 2 + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFOPPlus + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + _col2 + + + _col3 + + + _col4 + + + _col5 + + + _col6 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -875,18 +697,96 @@ - + + + _c0 + - key + _col0 - - src + + + + + + + + _c1 + + + _col1 + - + + + + + _c2 + + + _col2 + + + + + + + + + + _c3 + + + _col3 + + + + + + + + + + _c4 + + + _col4 + + + + + + + + + + _c5 + + + _col5 + + + + + + + + + + _c6 + + + _col6 + + + + + + @@ -910,7 +810,11 @@ src - + + + string + + @@ -967,7 +871,17 @@ - + + + key + + + src + + + + + @@ -1076,7 +990,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1088,7 +1002,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1145,11 +1059,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011470 + 1290531098 Index: ql/src/test/results/compiler/plan/input1.q.xml =================================================================== --- ql/src/test/results/compiler/plan/input1.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/input1.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011555 + 1290531121 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011555 + 1290531121 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 @@ -521,11 +521,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011553 + 1290531120 @@ -566,260 +566,143 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-32-35_827_7719579997360780269/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - src - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10002 - - _col0 - - - key - - - src - - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-01_558_953944160245547098/-ext-10000/ + + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - src - - - - - - - - - - _col1 - - - src - - - - - - - - - + - - - - - - - - - - key - - - src - - - - - - - - - - - - int - - - - - 100 - - - - + + + + _col1 + + + value - - + + src - - - boolean - - + + + _col0 + + + key + + + src + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -833,9 +716,9 @@ - + - key + _col0 src @@ -846,9 +729,9 @@ - + - value + _col1 src @@ -888,7 +771,11 @@ - + + + int + + 100 @@ -901,7 +788,11 @@ - + + + boolean + + @@ -938,10 +829,30 @@ - + + + key + + + src + + + + + - + + + value + + + src + + + + + @@ -1043,7 +954,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1055,7 +966,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1112,11 +1023,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011553 + 1290531120 Index: ql/src/test/results/compiler/plan/input4.q.xml =================================================================== --- ql/src/test/results/compiler/plan/input4.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/input4.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -22,7 +22,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-26_703_3879813904148546111/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-14_121_6158168614172964652/-ext-10000/ @@ -58,7 +58,7 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-26_703_3879813904148546111/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-14_121_6158168614172964652/-ext-10000 @@ -111,11 +111,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011606 + 1290531133 @@ -125,7 +125,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-26_703_3879813904148546111/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-14_121_6158168614172964652/-ext-10001 @@ -196,11 +196,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011603 + 1290531133 @@ -245,160 +245,274 @@ - - - - - _col1 - - - _col1 - - - - - string + + + + + + + + + _col1 + + + _col1 + + + + + string + + + + + + _col0 + + + _col0 + + + + + + - - - - _col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 + + + + - - + + + + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + reducesinkkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + + + reducesinkkey0 + + + + + + + _col0 + + + _col1 + + + + + + + + + _col0 + + + + + + + + + + -1 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - reducesinkkey0 + + + + CNTR_NAME_RS_3_NUM_INPUT_ROWS - - serialization.sort.order - + + + CNTR_NAME_RS_3_NUM_OUTPUT_ROWS - - columns.types - string + + CNTR_NAME_RS_3_TIME_TAKEN + + CNTR_NAME_RS_3_FATAL_ERROR + - - - - 1 - - - -1 - - - - - reducesinkkey0 + + RS_3 - - - - - - _col0 + + + + + + - - _col1 + + + + + + + + tkey + + + _col0 + + + + + + + + + + tvalue + + + _col1 + + + + + + + + + - - - - - - _col0 + + + + + + + + + + + + _col0 + + + + + - - + + + + + + int + + + + + 100 + + - - - - -1 - - - - - + + - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 + + + + boolean - - columns.types - string,string - - - escape.delim - \ - @@ -408,21 +522,21 @@ - CNTR_NAME_RS_3_NUM_INPUT_ROWS + CNTR_NAME_FIL_8_NUM_INPUT_ROWS - CNTR_NAME_RS_3_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_8_NUM_OUTPUT_ROWS - CNTR_NAME_RS_3_TIME_TAKEN + CNTR_NAME_FIL_8_TIME_TAKEN - CNTR_NAME_RS_3_FATAL_ERROR + CNTR_NAME_FIL_8_FATAL_ERROR - RS_3 + FIL_8 @@ -434,34 +548,7 @@ - - - - - tkey - - - _col0 - - - - - - - - - - tvalue - - - _col1 - - - - - - - + @@ -856,7 +943,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src tmap:src @@ -868,7 +955,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -925,11 +1012,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011603 + 1290531133 @@ -964,183 +1051,59 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-26_703_3879813904148546111/-ext-10000 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-26_703_3879813904148546111/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_7_NUM_INPUT_ROWS - - - CNTR_NAME_FS_7_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_7_TIME_TAKEN - - - CNTR_NAME_FS_7_FATAL_ERROR - - - - - FS_7 - - - - - - - - - - - - - - - - key - - - - - - - - - - - - - value - - - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - _col1 - - - tmap - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-14_121_6158168614172964652/-ext-10000 - - _col0 - - - _col0 - - - tmap - - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-14_121_6158168614172964652/-ext-10000/ + + + + + 1 + - CNTR_NAME_SEL_6_NUM_INPUT_ROWS + CNTR_NAME_FS_7_NUM_INPUT_ROWS - CNTR_NAME_SEL_6_NUM_OUTPUT_ROWS + CNTR_NAME_FS_7_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_6_TIME_TAKEN + CNTR_NAME_FS_7_TIME_TAKEN - CNTR_NAME_SEL_6_FATAL_ERROR + CNTR_NAME_FS_7_FATAL_ERROR - SEL_6 + FS_7 - + @@ -1151,10 +1114,10 @@ - _col0 + key - tmap + @@ -1164,10 +1127,10 @@ - _col1 + value - tmap + @@ -1182,73 +1145,80 @@ - - - - - - - - - - _col0 - - - tmap - - - - - - - - - - - - int - - - - - 100 - - - - + + + + _col1 + + + _col1 - - + + tmap - - - boolean - - + + + _col0 + + + _col0 + + + tmap + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_5_NUM_INPUT_ROWS + CNTR_NAME_SEL_6_NUM_INPUT_ROWS - CNTR_NAME_FIL_5_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_6_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_5_TIME_TAKEN + CNTR_NAME_SEL_6_TIME_TAKEN - CNTR_NAME_FIL_5_FATAL_ERROR + CNTR_NAME_SEL_6_FATAL_ERROR - FIL_5 + SEL_6 @@ -1262,7 +1232,7 @@ - + _col0 @@ -1275,7 +1245,7 @@ - + _col1 @@ -1343,10 +1313,30 @@ - + + + _col0 + + + tmap + + + + + - + + + _col1 + + + tmap + + + + + Index: ql/src/test/results/compiler/plan/input6.q.xml =================================================================== --- ql/src/test/results/compiler/plan/input6.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/input6.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011632 + 1290531139 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011632 + 1290531139 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 @@ -521,11 +521,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 transient_lastDdlTime - 1290011630 + 1290531139 @@ -566,246 +566,143 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-33-52_476_9052155505885239517/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - src1 - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10002 - - _col0 - - - key - - - src1 - - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-19_940_4056950164712671365/-ext-10000/ + + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - src1 - - - - - - - - - - _col1 - - - src1 - - - - - - - - - + - - - - - - - - - - key - - - src1 - - - - - - - + + + + _col1 + + + value - - + + src1 - - - boolean - - + + + _col0 + + + key + + + src1 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -819,9 +716,9 @@ - + - key + _col0 src1 @@ -832,9 +729,9 @@ - + - value + _col1 src1 @@ -877,7 +774,11 @@ - + + + boolean + + @@ -914,10 +815,30 @@ - + + + key + + + src1 + + + + + - + + + value + + + src1 + + + + + @@ -1019,7 +940,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 src1 @@ -1031,7 +952,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 src1 @@ -1088,11 +1009,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 transient_lastDdlTime - 1290011630 + 1290531139 Index: ql/src/test/results/compiler/plan/input9.q.xml =================================================================== --- ql/src/test/results/compiler/plan/input9.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/input9.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011672 + 1290531148 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011672 + 1290531148 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 @@ -521,11 +521,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 transient_lastDdlTime - 1290011670 + 1290531147 @@ -566,242 +566,137 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-34-32_553_569820262693254126/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - key - - - src1 - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10002 - - _col0 - - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-52-28_208_1025691484401950873/-ext-10000/ + + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - - - void - - - - - - - - - _col1 - - - src1 - - - - - - - - - + - - - - - - - - - - - - - - - - - - - - - + + + + _col1 + + + key - - + + src1 - - - boolean - - + + + _col0 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -815,10 +710,24 @@ - + - key + _col0 + + + + void + + + + + + + + + _col1 + src1 @@ -861,7 +770,11 @@ - + + + boolean + + @@ -898,7 +811,17 @@ - + + + key + + + src1 + + + + + @@ -1010,7 +933,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 src1 @@ -1022,7 +945,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 src1 @@ -1079,11 +1002,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src1 transient_lastDdlTime - 1290011670 + 1290531147 Index: ql/src/test/results/compiler/plan/input_part1.q.xml =================================================================== --- ql/src/test/results/compiler/plan/input_part1.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/input_part1.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -75,11 +75,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart transient_lastDdlTime - 1290011674 + 1290531148 @@ -120,279 +120,79 @@ - + - - - - - - - - - file:/tmp/njain/hive_2010-11-17_08-34-46_008_2039338248727502656/-ext-10001 - - - 1 - - - file:/tmp/njain/hive_2010-11-17_08-34-46_008_2039338248727502656/-ext-10001/ - - - - - org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - org.apache.hadoop.mapred.TextInputFormat - - - org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - - columns - _col0,_col1,_col2,_col3 - - - serialization.format - 1 - - - columns.types - string:string:string:string - - - - - - - 1 - - + + + + + file:/tmp/amarsri/hive_2010-11-23_08-52-31_041_3333310699766809193/-ext-10001 + + + 1 + + + file:/tmp/amarsri/hive_2010-11-23_08-52-31_041_3333310699766809193/-ext-10001/ + + + + + org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - + + org.apache.hadoop.mapred.TextInputFormat - - FS_3 + + org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - + + + + columns + _col0,_col1,_col2,_col3 - - - - - - - - - - _col0 - - - - - - - - string - - - - - - - - - _col1 - - - - - - - - - - - - - _col2 - - - - - - - - - - - - - _col3 - - - - - - - - - - + + serialization.format + 1 + + columns.types + string:string:string:string + - - - - - - _col3 - - - ds - - - true - - - srcpart - - - - - + + 1 - - _col2 - - - hr - - - true - - - srcpart - - - - - - - - _col1 - - - value - - - srcpart - - - - - - - - _col0 - - - key - - - srcpart - - - - - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - _col1 - - - _col2 - - - _col3 - - - - - - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + @@ -402,30 +202,28 @@ - - key - _col0 - srcpart + - + + + string + + - - value - _col1 - srcpart + @@ -434,14 +232,11 @@ - - hr - _col2 - srcpart + @@ -450,14 +245,11 @@ - - ds - _col3 - srcpart + @@ -472,181 +264,126 @@ - - - - - - - - - - - - - - - - - - key - - - srcpart - - - - - - - - - - - - int - - - - - 100 - - - - - - - - - - - - boolean - - - - - - - - - - - - - ds - - - true - - - srcpart - - - - - - - - - - - - - 2008-04-08 - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - hr - - - true - - - srcpart - - - - - - - - - - - - - 12 - - - - - - - - - - - - - - + + + + _col3 + + + ds - - + + true + + srcpart + - + + + _col2 + + + hr + + + true + + + srcpart + + + + + + + + _col1 + + + value + + + srcpart + + + + + + + + _col0 + + + key + + + srcpart + + + + + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + _col2 + + + _col3 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -660,10 +397,13 @@ - - + + key + + _col0 + srcpart @@ -673,10 +413,13 @@ - - + + value + + _col1 + srcpart @@ -686,9 +429,12 @@ - + + + hr + - ds + _col2 srcpart @@ -699,9 +445,12 @@ - + + + ds + - hr + _col3 srcpart @@ -741,7 +490,11 @@ - + + + int + + 100 @@ -754,7 +507,11 @@ - + + + boolean + + @@ -791,16 +548,56 @@ - + + + key + + + srcpart + + + + + - + + + value + + + srcpart + + + + + - + + + ds + + + srcpart + + + + + - + + + hr + + + srcpart + + + + + @@ -899,7 +696,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 srcpart @@ -911,7 +708,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 hr=12 @@ -981,11 +778,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart transient_lastDdlTime - 1290011674 + 1290531148 Index: ql/src/test/results/compiler/plan/input_testxpath2.q.xml =================================================================== --- ql/src/test/results/compiler/plan/input_testxpath2.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/input_testxpath2.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -66,11 +66,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src_thrift + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src_thrift transient_lastDdlTime - 1290011726 + 1290531159 @@ -111,328 +111,79 @@ - + - - - - - - - - - file:/tmp/njain/hive_2010-11-17_08-35-27_510_5379644232449949052/-ext-10001 - - - 1 - - - file:/tmp/njain/hive_2010-11-17_08-35-27_510_5379644232449949052/-ext-10001/ - - - - - org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - org.apache.hadoop.mapred.TextInputFormat - - - org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - - columns - _col0,_col1,_col2 - - - serialization.format - 1 - - - columns.types - int:int:int - - - - - - - 1 - - + + + + + file:/tmp/amarsri/hive_2010-11-23_08-52-39_307_2419749432070823713/-ext-10001 + + + 1 + + + file:/tmp/amarsri/hive_2010-11-23_08-52-39_307_2419749432070823713/-ext-10001/ + + + + + org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - + + org.apache.hadoop.mapred.TextInputFormat - - FS_3 + + org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - + + + + columns + _col0,_col1,_col2 - - - - - - - - - - _col0 - - - - - - - - int - - - - - - - - - _col1 - - - - - - - - - - - - - _col2 - - - - - - - - - - + + serialization.format + 1 - - - - - - - - - - _col2 - - - - - - - mstringstring - - - src_thrift - - - - - - - string - - - - - - - - - + + columns.types + int:int:int - - - - - - - - _col1 - - - - - - - lintstring - - - src_thrift - - - - - - - - - myint - - - mystring - - - underscore_int - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - + + 1 - - _col0 - - - - - - - lint - - - src_thrift - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - _col1 - - - _col2 - - - - - - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + @@ -442,25 +193,29 @@ - - _c0 - _col0 + + + - + + + int + + - - _c1 - _col1 + + + @@ -468,12 +223,12 @@ - - _c2 - _col2 + + + @@ -487,114 +242,188 @@ - - - - + + + + _col2 + - - - - - - - lint - - - src_thrift - - - - - - - + + + mstringstring - - + + src_thrift - - - boolean + + + + + string + + + + + + + + + + + + + + + + + _col1 + + + - - - - - - + + + lintstring + + + src_thrift + + + + + + - - - mstringstring - - - src_thrift - - - - - + myint + + mystring + + + underscore_int + - - + + + + + + + + + + + + - - - - - + + + + + + + + + + + + + + _col0 + + + + + + + lint + + src_thrift + - + + + + + - + - + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + _col2 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -608,41 +437,41 @@ - + + + _c0 + - lint + _col0 - - src_thrift - - + - + + + _c1 + - lintstring + _col1 - - src_thrift - - + - + + + _c2 + - mstringstring + _col2 - - src_thrift - - + @@ -683,7 +512,11 @@ - + + + boolean + + @@ -797,7 +630,17 @@ - + + + lint + + + src_thrift + + + + + @@ -817,10 +660,30 @@ - + + + lintstring + + + src_thrift + + + + + - + + + mstringstring + + + src_thrift + + + + + @@ -922,7 +785,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src_thrift + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src_thrift src_thrift @@ -934,7 +797,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src_thrift + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src_thrift src_thrift @@ -995,11 +858,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src_thrift + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src_thrift transient_lastDdlTime - 1290011726 + 1290531159 Index: ql/src/test/results/compiler/plan/join4.q.xml =================================================================== --- ql/src/test/results/compiler/plan/join4.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/join4.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011781 + 1290531170 @@ -149,11 +149,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011781 + 1290531170 @@ -194,249 +194,34 @@ - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - string - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_6_NUM_INPUT_ROWS - - - CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_6_TIME_TAKEN - - - CNTR_NAME_RS_6_FATAL_ERROR - - - - - RS_6 - - - - - - - - - - - - - - - - VALUE._col0 - - - a - - - - - - - - - - VALUE._col1 - - - a - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src1 - - + + + string + + - _col0 - + VALUE._col0 + - key + _col0 - - src1 - @@ -445,18 +230,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -466,31 +300,73 @@ + + + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_5_NUM_INPUT_ROWS + CNTR_NAME_RS_6_NUM_INPUT_ROWS - CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_5_TIME_TAKEN + CNTR_NAME_RS_6_TIME_TAKEN - CNTR_NAME_SEL_5_FATAL_ERROR + CNTR_NAME_RS_6_FATAL_ERROR - SEL_5 + RS_6 - + @@ -501,8 +377,11 @@ - _col0 + VALUE._col0 + + a + @@ -511,8 +390,11 @@ - _col1 + VALUE._col1 + + a + @@ -526,124 +408,80 @@ - - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - int - - - - - 10 - - - - - - - - - - - - boolean - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - - 20 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src1 - + + + _col0 + + + key + + + src1 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_SEL_5_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_SEL_5_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_SEL_5_FATAL_ERROR - FIL_4 + SEL_5 @@ -657,26 +495,20 @@ - + - key + _col0 - - src1 - - + - value + _col1 - - src1 - @@ -716,7 +548,11 @@ - + + + int + + 10 @@ -729,7 +565,11 @@ - + + + boolean + + @@ -813,10 +653,30 @@ - + + + key + + + src1 + + + + + - + + + value + + + src1 + + + + + @@ -916,252 +776,34 @@ - + - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - 1 - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_7_NUM_INPUT_ROWS - - - CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_7_TIME_TAKEN - - - CNTR_NAME_RS_7_FATAL_ERROR - - - - - RS_7 - - - - - - - - - - - - - - - - VALUE._col0 - - - b - - - - - - - - - - VALUE._col1 - - - b - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src2 - - _col0 - + VALUE._col0 + - key + _col0 - - src2 - @@ -1170,18 +812,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -1191,31 +882,76 @@ + + + + + 1 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_RS_7_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_RS_7_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_RS_7_FATAL_ERROR - SEL_2 + RS_7 - + @@ -1226,8 +962,11 @@ - _col0 + VALUE._col0 + + b + @@ -1236,8 +975,11 @@ - _col1 + VALUE._col1 + + b + @@ -1251,121 +993,85 @@ - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 15 - - - - - - - - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 25 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src2 - + + + _col0 + + + key + + + src2 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 - + @@ -1374,26 +1080,20 @@ - + - key + _col0 - - src2 - - + - value + _col1 - - src2 - @@ -1530,10 +1230,30 @@ - + + + key + + + src2 + + + + + - + + + value + + + src2 + + + + + @@ -1637,7 +1357,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src c:a:src1 @@ -1652,7 +1372,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1709,11 +1429,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011781 + 1290531170 @@ -1760,13 +1480,13 @@ - file:/tmp/njain/hive_2010-11-17_08-36-25_040_3628580690718411176/-ext-10001 + file:/tmp/amarsri/hive_2010-11-23_08-52-50_931_6163419363689703256/-ext-10001 1 - file:/tmp/njain/hive_2010-11-17_08-36-25_040_3628580690718411176/-ext-10001/ + file:/tmp/amarsri/hive_2010-11-23_08-52-50_931_6163419363689703256/-ext-10001/ Index: ql/src/test/results/compiler/plan/join5.q.xml =================================================================== --- ql/src/test/results/compiler/plan/join5.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/join5.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011796 + 1290531173 @@ -149,11 +149,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011796 + 1290531173 @@ -194,249 +194,34 @@ - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - string - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_6_NUM_INPUT_ROWS - - - CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_6_TIME_TAKEN - - - CNTR_NAME_RS_6_FATAL_ERROR - - - - - RS_6 - - - - - - - - - - - - - - - - VALUE._col0 - - - a - - - - - - - - - - VALUE._col1 - - - a - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src1 - - + + + string + + - _col0 - + VALUE._col0 + - key + _col0 - - src1 - @@ -445,18 +230,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -466,31 +300,73 @@ + + + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_5_NUM_INPUT_ROWS + CNTR_NAME_RS_6_NUM_INPUT_ROWS - CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_5_TIME_TAKEN + CNTR_NAME_RS_6_TIME_TAKEN - CNTR_NAME_SEL_5_FATAL_ERROR + CNTR_NAME_RS_6_FATAL_ERROR - SEL_5 + RS_6 - + @@ -501,8 +377,11 @@ - _col0 + VALUE._col0 + + a + @@ -511,8 +390,11 @@ - _col1 + VALUE._col1 + + a + @@ -526,124 +408,80 @@ - - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - int - - - - - 10 - - - - - - - - - - - - boolean - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - - 20 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src1 - + + + _col0 + + + key + + + src1 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_SEL_5_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_SEL_5_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_SEL_5_FATAL_ERROR - FIL_4 + SEL_5 @@ -657,26 +495,20 @@ - + - key + _col0 - - src1 - - + - value + _col1 - - src1 - @@ -716,7 +548,11 @@ - + + + int + + 10 @@ -729,7 +565,11 @@ - + + + boolean + + @@ -813,10 +653,30 @@ - + + + key + + + src1 + + + + + - + + + value + + + src1 + + + + + @@ -916,252 +776,34 @@ - + - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - 1 - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_7_NUM_INPUT_ROWS - - - CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_7_TIME_TAKEN - - - CNTR_NAME_RS_7_FATAL_ERROR - - - - - RS_7 - - - - - - - - - - - - - - - - VALUE._col0 - - - b - - - - - - - - - - VALUE._col1 - - - b - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src2 - - _col0 - + VALUE._col0 + - key + _col0 - - src2 - @@ -1170,18 +812,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -1191,31 +882,76 @@ + + + + + 1 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_RS_7_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_RS_7_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_RS_7_FATAL_ERROR - SEL_2 + RS_7 - + @@ -1226,8 +962,11 @@ - _col0 + VALUE._col0 + + b + @@ -1236,8 +975,11 @@ - _col1 + VALUE._col1 + + b + @@ -1251,121 +993,85 @@ - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 15 - - - - - - - - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 25 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src2 - + + + _col0 + + + key + + + src2 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 - + @@ -1374,26 +1080,20 @@ - + - key + _col0 - - src2 - - + - value + _col1 - - src2 - @@ -1530,10 +1230,30 @@ - + + + key + + + src2 + + + + + - + + + value + + + src2 + + + + + @@ -1637,7 +1357,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src c:a:src1 @@ -1652,7 +1372,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1709,11 +1429,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011796 + 1290531173 @@ -1760,13 +1480,13 @@ - file:/tmp/njain/hive_2010-11-17_08-36-39_583_6998724680726663672/-ext-10001 + file:/tmp/amarsri/hive_2010-11-23_08-52-53_917_3182737899161061784/-ext-10001 1 - file:/tmp/njain/hive_2010-11-17_08-36-39_583_6998724680726663672/-ext-10001/ + file:/tmp/amarsri/hive_2010-11-23_08-52-53_917_3182737899161061784/-ext-10001/ Index: ql/src/test/results/compiler/plan/join6.q.xml =================================================================== --- ql/src/test/results/compiler/plan/join6.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/join6.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011810 + 1290531176 @@ -149,11 +149,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011810 + 1290531176 @@ -194,249 +194,34 @@ - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - string - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_6_NUM_INPUT_ROWS - - - CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_6_TIME_TAKEN - - - CNTR_NAME_RS_6_FATAL_ERROR - - - - - RS_6 - - - - - - - - - - - - - - - - VALUE._col0 - - - a - - - - - - - - - - VALUE._col1 - - - a - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src1 - - + + + string + + - _col0 - + VALUE._col0 + - key + _col0 - - src1 - @@ -445,18 +230,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -466,31 +300,73 @@ + + + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_5_NUM_INPUT_ROWS + CNTR_NAME_RS_6_NUM_INPUT_ROWS - CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_5_TIME_TAKEN + CNTR_NAME_RS_6_TIME_TAKEN - CNTR_NAME_SEL_5_FATAL_ERROR + CNTR_NAME_RS_6_FATAL_ERROR - SEL_5 + RS_6 - + @@ -501,8 +377,11 @@ - _col0 + VALUE._col0 + + a + @@ -511,8 +390,11 @@ - _col1 + VALUE._col1 + + a + @@ -526,124 +408,80 @@ - - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - int - - - - - 10 - - - - - - - - - - - - boolean - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - - 20 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src1 - + + + _col0 + + + key + + + src1 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_SEL_5_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_SEL_5_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_SEL_5_FATAL_ERROR - FIL_4 + SEL_5 @@ -657,26 +495,20 @@ - + - key + _col0 - - src1 - - + - value + _col1 - - src1 - @@ -716,7 +548,11 @@ - + + + int + + 10 @@ -729,7 +565,11 @@ - + + + boolean + + @@ -813,10 +653,30 @@ - + + + key + + + src1 + + + + + - + + + value + + + src1 + + + + + @@ -916,252 +776,34 @@ - + - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - 1 - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_7_NUM_INPUT_ROWS - - - CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_7_TIME_TAKEN - - - CNTR_NAME_RS_7_FATAL_ERROR - - - - - RS_7 - - - - - - - - - - - - - - - - VALUE._col0 - - - b - - - - - - - - - - VALUE._col1 - - - b - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src2 - - _col0 - + VALUE._col0 + - key + _col0 - - src2 - @@ -1170,18 +812,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -1191,31 +882,76 @@ + + + + + 1 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_RS_7_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_RS_7_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_RS_7_FATAL_ERROR - SEL_2 + RS_7 - + @@ -1226,8 +962,11 @@ - _col0 + VALUE._col0 + + b + @@ -1236,8 +975,11 @@ - _col1 + VALUE._col1 + + b + @@ -1251,121 +993,85 @@ - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 15 - - - - - - - - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 25 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src2 - + + + _col0 + + + key + + + src2 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 - + @@ -1374,26 +1080,20 @@ - + - key + _col0 - - src2 - - + - value + _col1 - - src2 - @@ -1530,10 +1230,30 @@ - + + + key + + + src2 + + + + + - + + + value + + + src2 + + + + + @@ -1637,7 +1357,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src c:a:src1 @@ -1652,7 +1372,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1709,11 +1429,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011810 + 1290531176 @@ -1760,13 +1480,13 @@ - file:/tmp/njain/hive_2010-11-17_08-36-54_403_2015400014988768548/-ext-10001 + file:/tmp/amarsri/hive_2010-11-23_08-52-56_843_3634496630516998560/-ext-10001 1 - file:/tmp/njain/hive_2010-11-17_08-36-54_403_2015400014988768548/-ext-10001/ + file:/tmp/amarsri/hive_2010-11-23_08-52-56_843_3634496630516998560/-ext-10001/ Index: ql/src/test/results/compiler/plan/join7.q.xml =================================================================== --- ql/src/test/results/compiler/plan/join7.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/join7.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011825 + 1290531179 @@ -149,11 +149,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011825 + 1290531179 @@ -236,11 +236,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011825 + 1290531179 @@ -281,249 +281,34 @@ - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - string - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_9_NUM_INPUT_ROWS - - - CNTR_NAME_RS_9_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_9_TIME_TAKEN - - - CNTR_NAME_RS_9_FATAL_ERROR - - - - - RS_9 - - - - - - - - - - - - - - - - VALUE._col0 - - - a - - - - - - - - - - VALUE._col1 - - - a - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src1 - - + + + string + + - _col0 - + VALUE._col0 + - key + _col0 - - src1 - @@ -532,18 +317,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -553,31 +387,73 @@ + + + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_8_NUM_INPUT_ROWS + CNTR_NAME_RS_9_NUM_INPUT_ROWS - CNTR_NAME_SEL_8_NUM_OUTPUT_ROWS + CNTR_NAME_RS_9_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_8_TIME_TAKEN + CNTR_NAME_RS_9_TIME_TAKEN - CNTR_NAME_SEL_8_FATAL_ERROR + CNTR_NAME_RS_9_FATAL_ERROR - SEL_8 + RS_9 - + @@ -588,8 +464,11 @@ - _col0 + VALUE._col0 + + a + @@ -598,8 +477,11 @@ - _col1 + VALUE._col1 + + a + @@ -613,124 +495,80 @@ - - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - int - - - - - 10 - - - - - - - - - - - - boolean - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - - 20 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src1 - + + + _col0 + + + key + + + src1 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_7_NUM_INPUT_ROWS + CNTR_NAME_SEL_8_NUM_INPUT_ROWS - CNTR_NAME_FIL_7_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_8_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_7_TIME_TAKEN + CNTR_NAME_SEL_8_TIME_TAKEN - CNTR_NAME_FIL_7_FATAL_ERROR + CNTR_NAME_SEL_8_FATAL_ERROR - FIL_7 + SEL_8 @@ -744,26 +582,20 @@ - + - key + _col0 - - src1 - - + - value + _col1 - - src1 - @@ -803,7 +635,11 @@ - + + + int + + 10 @@ -816,7 +652,11 @@ - + + + boolean + + @@ -900,10 +740,30 @@ - + + + key + + + src1 + + + + + - + + + value + + + src1 + + + + + @@ -1003,252 +863,34 @@ - + - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - 1 - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_10_NUM_INPUT_ROWS - - - CNTR_NAME_RS_10_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_10_TIME_TAKEN - - - CNTR_NAME_RS_10_FATAL_ERROR - - - - - RS_10 - - - - - - - - - - - - - - - - VALUE._col0 - - - b - - - - - - - - - - VALUE._col1 - - - b - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src2 - - _col0 - + VALUE._col0 + - key + _col0 - - src2 - @@ -1257,18 +899,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -1278,31 +969,76 @@ + + + + + 1 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_RS_10_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_RS_10_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_RS_10_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_RS_10_FATAL_ERROR - SEL_2 + RS_10 - + @@ -1313,8 +1049,11 @@ - _col0 + VALUE._col0 + + b + @@ -1323,8 +1062,11 @@ - _col1 + VALUE._col1 + + b + @@ -1338,121 +1080,85 @@ - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 15 - - - - - - - - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 25 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src2 - + + + _col0 + + + key + + + src2 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 - + @@ -1461,26 +1167,20 @@ - + - key + _col0 - - src2 - - + - value + _col1 - - src2 - @@ -1617,10 +1317,30 @@ - + + + key + + + src2 + + + + + - + + + value + + + src2 + + + + + @@ -1716,252 +1436,34 @@ - + - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - 2 - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_11_NUM_INPUT_ROWS - - - CNTR_NAME_RS_11_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_11_TIME_TAKEN - - - CNTR_NAME_RS_11_FATAL_ERROR - - - - - RS_11 - - - - - - - - - - - - - - - - VALUE._col0 - - - c - - - - - - - - - - VALUE._col1 - - - c - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src3 - - _col0 - + VALUE._col0 + - key + _col0 - - src3 - @@ -1970,18 +1472,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -1991,31 +1542,76 @@ + + + + + 2 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_5_NUM_INPUT_ROWS + CNTR_NAME_RS_11_NUM_INPUT_ROWS - CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + CNTR_NAME_RS_11_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_5_TIME_TAKEN + CNTR_NAME_RS_11_TIME_TAKEN - CNTR_NAME_SEL_5_FATAL_ERROR + CNTR_NAME_RS_11_FATAL_ERROR - SEL_5 + RS_11 - + @@ -2026,8 +1622,11 @@ - _col0 + VALUE._col0 + + c + @@ -2036,8 +1635,11 @@ - _col1 + VALUE._col1 + + c + @@ -2051,121 +1653,85 @@ - - - - - - - - - - - - - - key - - - src3 - - - - - - - - - - - - - 20 - - - - - - - - - - - - - - - - - - - - - key - - - src3 - - - - - - - - - - - - - 25 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src3 - + + + _col0 + + + key + + + src3 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_SEL_5_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_SEL_5_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_SEL_5_FATAL_ERROR - FIL_4 + SEL_5 - + @@ -2174,26 +1740,20 @@ - + - key + _col0 - - src3 - - + - value + _col1 - - src3 - @@ -2330,10 +1890,30 @@ - + + + key + + + src3 + + + + + - + + + value + + + src3 + + + + + @@ -2437,7 +2017,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src c:a:src1 @@ -2455,7 +2035,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -2512,11 +2092,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011825 + 1290531179 @@ -2563,13 +2143,13 @@ - file:/tmp/njain/hive_2010-11-17_08-37-10_192_6970711124262320892/-ext-10001 + file:/tmp/amarsri/hive_2010-11-23_08-52-59_810_7762362077228124174/-ext-10001 1 - file:/tmp/njain/hive_2010-11-17_08-37-10_192_6970711124262320892/-ext-10001/ + file:/tmp/amarsri/hive_2010-11-23_08-52-59_810_7762362077228124174/-ext-10001/ Index: ql/src/test/results/compiler/plan/join8.q.xml =================================================================== --- ql/src/test/results/compiler/plan/join8.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/join8.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011841 + 1290531182 @@ -149,11 +149,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011841 + 1290531182 @@ -194,249 +194,34 @@ - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - string - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_6_NUM_INPUT_ROWS - - - CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_6_TIME_TAKEN - - - CNTR_NAME_RS_6_FATAL_ERROR - - - - - RS_6 - - - - - - - - - - - - - - - - VALUE._col0 - - - a - - - - - - - - - - VALUE._col1 - - - a - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src1 - - + + + string + + - _col0 - + VALUE._col0 + - key + _col0 - - src1 - @@ -445,18 +230,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -466,31 +300,73 @@ + + + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_5_NUM_INPUT_ROWS + CNTR_NAME_RS_6_NUM_INPUT_ROWS - CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + CNTR_NAME_RS_6_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_5_TIME_TAKEN + CNTR_NAME_RS_6_TIME_TAKEN - CNTR_NAME_SEL_5_FATAL_ERROR + CNTR_NAME_RS_6_FATAL_ERROR - SEL_5 + RS_6 - + @@ -501,8 +377,11 @@ - _col0 + VALUE._col0 + + a + @@ -511,8 +390,11 @@ - _col1 + VALUE._col1 + + a + @@ -526,124 +408,80 @@ - - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - int - - - - - 10 - - - - - - - - - - - - boolean - - - - - - - - - - - - - key - - - src1 - - - - - - - - - - - - - 20 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src1 - + + + _col0 + + + key + + + src1 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_SEL_5_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_SEL_5_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_SEL_5_FATAL_ERROR - FIL_4 + SEL_5 @@ -657,26 +495,20 @@ - + - key + _col0 - - src1 - - + - value + _col1 - - src1 - @@ -716,7 +548,11 @@ - + + + int + + 10 @@ -729,7 +565,11 @@ - + + + boolean + + @@ -785,21 +625,21 @@ - CNTR_NAME_FIL_13_NUM_INPUT_ROWS + CNTR_NAME_FIL_14_NUM_INPUT_ROWS - CNTR_NAME_FIL_13_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_14_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_13_TIME_TAKEN + CNTR_NAME_FIL_14_TIME_TAKEN - CNTR_NAME_FIL_13_FATAL_ERROR + CNTR_NAME_FIL_14_FATAL_ERROR - FIL_13 + FIL_14 @@ -813,10 +653,30 @@ - + + + key + + + src1 + + + + + - + + + value + + + src1 + + + + + @@ -916,252 +776,34 @@ - + - + - - - - - - - - - VALUE._col1 - - - _col1 - - - - - - - - VALUE._col0 - - - _col0 - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - joinkey0 - - - serialization.sort.order - + - - - columns.types - string - - - - - - - 1 - - - -1 - - - - - reducesinkkey0 - - - - - - - _col0 - - - _col1 - - - - - - - - 1 - - - - - - - - - - - - - - - org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - - org.apache.hadoop.mapred.SequenceFileInputFormat - - - org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - - - - columns - _col0,_col1 - - - columns.types - string,string - - - escape.delim - \ - - - - - - - - - - - CNTR_NAME_RS_7_NUM_INPUT_ROWS - - - CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - - - CNTR_NAME_RS_7_TIME_TAKEN - - - CNTR_NAME_RS_7_FATAL_ERROR - - - - - RS_7 - - - - - - - - - - - - - - - - VALUE._col0 - - - b - - - - - - - - - - VALUE._col1 - - - b - - - - - - - - - - - - - - + - _col1 - + VALUE._col1 + - value + _col1 - - src2 - - _col0 - + VALUE._col0 + - key + _col0 - - src2 - @@ -1170,18 +812,67 @@ - - - + + + + + + - + + + _col0 + + + + + + + + + + + org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + joinkey0 + + + serialization.sort.order + + + + + columns.types + string + + + + + + + 1 + + + -1 + + + - + reducesinkkey0 - + _col0 @@ -1191,31 +882,76 @@ + + + + + 1 + + + + + + + + + + + + + + + org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + + org.apache.hadoop.mapred.SequenceFileInputFormat + + + org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + + + + + columns + _col0,_col1 + + + columns.types + string,string + + + escape.delim + \ + + + + + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_RS_7_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_RS_7_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_RS_7_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_RS_7_FATAL_ERROR - SEL_2 + RS_7 - + @@ -1226,8 +962,11 @@ - _col0 + VALUE._col0 + + b + @@ -1236,8 +975,11 @@ - _col1 + VALUE._col1 + + b + @@ -1251,121 +993,85 @@ - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 15 - - - - - - - - - - - - - - - - - - - - - key - - - src2 - - - - - - - - - - - - - 25 - - - - - - - - - - - - - - + + + + _col1 + + + value - - + + src2 - + + + _col0 + + + key + + + src2 + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 - + @@ -1374,26 +1080,20 @@ - + - key + _col0 - - src2 - - + - value + _col1 - - src2 - @@ -1502,21 +1202,21 @@ - CNTR_NAME_FIL_14_NUM_INPUT_ROWS + CNTR_NAME_FIL_15_NUM_INPUT_ROWS - CNTR_NAME_FIL_14_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_15_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_14_TIME_TAKEN + CNTR_NAME_FIL_15_TIME_TAKEN - CNTR_NAME_FIL_14_FATAL_ERROR + CNTR_NAME_FIL_15_FATAL_ERROR - FIL_14 + FIL_15 @@ -1530,10 +1230,30 @@ - + + + key + + + src2 + + + + + - + + + value + + + src2 + + + + + @@ -1637,7 +1357,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src c:a:src1 @@ -1652,7 +1372,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1709,11 +1429,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011841 + 1290531182 @@ -1748,11 +1468,11 @@ - + - + @@ -1764,13 +1484,13 @@ - file:/tmp/njain/hive_2010-11-17_08-37-25_411_1291497521396190225/-ext-10001 + file:/tmp/amarsri/hive_2010-11-23_08-53-03_062_6860367279508604896/-ext-10001 1 - file:/tmp/njain/hive_2010-11-17_08-37-25_411_1291497521396190225/-ext-10001/ + file:/tmp/amarsri/hive_2010-11-23_08-53-03_062_6860367279508604896/-ext-10001/ @@ -2002,7 +1722,7 @@ - + @@ -2082,95 +1802,125 @@ - - - - - - - - - - - - - - _col2 - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - - - - - - - - - - - - - - - + + + + _col3 + + + _col3 - - + + b - + + + _col2 + + + _col2 + + + b + + + + + + + + _col1 + + + _col1 + + + a + + + + + + + + _col0 + + + _col0 + + + a + + + + + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + _col2 + + + _col3 + + + + + - CNTR_NAME_FIL_10_NUM_INPUT_ROWS + CNTR_NAME_SEL_9_NUM_INPUT_ROWS - CNTR_NAME_FIL_10_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_9_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_10_TIME_TAKEN + CNTR_NAME_SEL_9_TIME_TAKEN - CNTR_NAME_FIL_10_FATAL_ERROR + CNTR_NAME_SEL_9_FATAL_ERROR - FIL_10 + SEL_9 - + @@ -2179,7 +1929,7 @@ - + _col0 @@ -2189,7 +1939,7 @@ - + _col1 @@ -2199,7 +1949,7 @@ - + _col2 @@ -2209,7 +1959,7 @@ - + _col3 @@ -2226,120 +1976,96 @@ - - - - _col3 - - - _col3 + + + + + + + + + + + + + + _col2 + + + b + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + a + + + + + + + + + + + + + + + + + - - b + + - + - - _col2 - - - _col2 - - - b - - - - - - - - _col1 - - - _col1 - - - a - - - - - - - - _col0 - - - _col0 - - - a - - - - - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - _col1 - - - _col2 - - - _col3 - - - - - - CNTR_NAME_SEL_9_NUM_INPUT_ROWS + CNTR_NAME_FIL_13_NUM_INPUT_ROWS - CNTR_NAME_SEL_9_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_13_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_9_TIME_TAKEN + CNTR_NAME_FIL_13_TIME_TAKEN - CNTR_NAME_SEL_9_FATAL_ERROR + CNTR_NAME_FIL_13_FATAL_ERROR - SEL_9 + FIL_13 @@ -2353,16 +2079,56 @@ - + + + _col0 + + + a + + + + + - + + + _col1 + + + a + + + + + - + + + _col2 + + + b + + + + + - + + + _col3 + + + b + + + + + @@ -2588,56 +2354,16 @@ - - - _col0 - - - a - - - - - + - - - _col1 - - - a - - - - - + - - - _col2 - - - b - - - - - + - - - _col3 - - - b - - - - - + Index: ql/src/test/results/compiler/plan/sample1.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample1.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample1.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -75,11 +75,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart transient_lastDdlTime - 1290011847 + 1290531183 @@ -120,282 +120,79 @@ - + - - - - - - - - - file:/tmp/njain/hive_2010-11-17_08-37-40_625_4189024771035527066/-ext-10001 - - - 1 - - - file:/tmp/njain/hive_2010-11-17_08-37-40_625_4189024771035527066/-ext-10001/ - - - - - org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - org.apache.hadoop.mapred.TextInputFormat - - - org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - - columns - _col0,_col1,_col2,_col3 - - - serialization.format - 1 - - - columns.types - string:string:string:string - - - - - - - 1 - - + + + + + file:/tmp/amarsri/hive_2010-11-23_08-53-06_114_7245877054668413999/-ext-10001 + + + 1 + + + file:/tmp/amarsri/hive_2010-11-23_08-53-06_114_7245877054668413999/-ext-10001/ + + + + + org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - - CNTR_NAME_FS_4_NUM_INPUT_ROWS - - - CNTR_NAME_FS_4_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_4_TIME_TAKEN - - - CNTR_NAME_FS_4_FATAL_ERROR - - + + org.apache.hadoop.mapred.TextInputFormat - - FS_4 + + org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - + + + + columns + _col0,_col1,_col2,_col3 - - - - - - - - - - _col0 - - - - - - - - string - - - - - - - - - _col1 - - - - - - - - - - - - - _col2 - - - - - - - - - - - - - _col3 - - - - - - - - - - + + serialization.format + 1 + + columns.types + string:string:string:string + - - - - - - _col3 - - - hr - - - true - - - s - - - - - + + 1 - - _col2 - - - ds - - - true - - - s - - - - - - - - _col1 - - - value - - - s - - - - - - - - _col0 - - - key - - - s - - - - - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - _col1 - - - _col2 - - - _col3 - - - - - true - - - - CNTR_NAME_SEL_3_NUM_INPUT_ROWS + CNTR_NAME_FS_4_NUM_INPUT_ROWS - CNTR_NAME_SEL_3_NUM_OUTPUT_ROWS + CNTR_NAME_FS_4_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_3_TIME_TAKEN + CNTR_NAME_FS_4_TIME_TAKEN - CNTR_NAME_SEL_3_FATAL_ERROR + CNTR_NAME_FS_4_FATAL_ERROR - SEL_3 + FS_4 - + @@ -405,30 +202,28 @@ - - key - _col0 - s + - + + + string + + - - value - _col1 - s + @@ -437,14 +232,11 @@ - - ds - _col2 - s + @@ -453,14 +245,11 @@ - - hr - _col3 - s + @@ -475,126 +264,129 @@ - - - - - - - - - - - - - - ds - - - true - - - s - - - - - - - - - - - - - 2008-04-08 - - - - - - - - - - - - boolean - - - - - - - - - - - - - hr - - - true - - - s - - - - - - - - - - - - - 11 - - - - - - - - - - - - - - + + + + _col3 + + + hr - - + + true + + s + - + + + _col2 + + + ds + + + true + + + s + + + + + + + + _col1 + + + value + + + s + + + + + + + + _col0 + + + key + + + s + + + + + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + _col2 + + + _col3 + + + + + true + + + - CNTR_NAME_FIL_2_NUM_INPUT_ROWS + CNTR_NAME_SEL_3_NUM_INPUT_ROWS - CNTR_NAME_FIL_2_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_3_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_2_TIME_TAKEN + CNTR_NAME_SEL_3_TIME_TAKEN - CNTR_NAME_FIL_2_FATAL_ERROR + CNTR_NAME_SEL_3_FATAL_ERROR - FIL_2 + SEL_3 @@ -608,10 +400,13 @@ - - + + key + + _col0 + s @@ -621,10 +416,13 @@ - - + + value + + _col1 + s @@ -634,10 +432,13 @@ - - + + ds + + _col2 + s @@ -647,10 +448,13 @@ - - + + hr + + _col3 + s @@ -718,7 +522,7 @@ - + int @@ -729,7 +533,7 @@ - + 2147483647 @@ -752,14 +556,14 @@ - + - + 1 @@ -782,14 +586,14 @@ - + - + 0 @@ -802,7 +606,11 @@ - + + + boolean + + @@ -839,16 +647,56 @@ - + + + key + + + s + + + + + - + + + value + + + s + + + + + - + + + ds + + + s + + + + + - + + + hr + + + s + + + + + @@ -960,7 +808,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 s @@ -972,7 +820,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 hr=11 @@ -1042,11 +890,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcpart + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcpart transient_lastDdlTime - 1290011847 + 1290531183 Index: ql/src/test/results/compiler/plan/sample2.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample2.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample2.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011875 + 1290531188 @@ -159,21 +159,21 @@ - CNTR_NAME_FS_6_NUM_INPUT_ROWS + CNTR_NAME_FS_5_NUM_INPUT_ROWS - CNTR_NAME_FS_6_NUM_OUTPUT_ROWS + CNTR_NAME_FS_5_NUM_OUTPUT_ROWS - CNTR_NAME_FS_6_TIME_TAKEN + CNTR_NAME_FS_5_TIME_TAKEN - CNTR_NAME_FS_6_FATAL_ERROR + CNTR_NAME_FS_5_FATAL_ERROR - FS_6 + FS_5 @@ -227,21 +227,21 @@ - CNTR_NAME_TS_5_NUM_INPUT_ROWS + CNTR_NAME_TS_4_NUM_INPUT_ROWS - CNTR_NAME_TS_5_NUM_OUTPUT_ROWS + CNTR_NAME_TS_4_NUM_OUTPUT_ROWS - CNTR_NAME_TS_5_TIME_TAKEN + CNTR_NAME_TS_4_TIME_TAKEN - CNTR_NAME_TS_5_FATAL_ERROR + CNTR_NAME_TS_4_FATAL_ERROR - TS_5 + TS_4 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011875 + 1290531188 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 @@ -525,11 +525,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011867 + 1290531187 @@ -570,351 +570,150 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-37-55_806_6771802665675897534/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - s - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10002 - - _col0 - - - key - - - s - - - - - int - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-08_976_8610659789811612021/-ext-10000/ - - true + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - s - - - - - - - - - - _col1 - - - s - - - - - - - - - + - - - - true + + + + _col1 + + + value + + + s + + + + + - - - - - - - - - - - - - - - - - - - - key - - - s - - - - - - - - - - - - - - - - - - - - - - - 2147483647 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPBitAnd - - - & - - - - - - - - - - - - - - - 2 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPMod - - - % - - - - - - - - - - - - - - - 0 - - - - + + _col0 + + + key - - + + s - + - boolean + int - - + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -928,9 +727,9 @@ - + - key + _col0 s @@ -941,9 +740,9 @@ - + - value + _col1 s @@ -963,6 +762,9 @@ + + true + @@ -1078,30 +880,37 @@ - + + + boolean + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_FIL_1_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_FIL_1_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_FIL_1_FATAL_ERROR - FIL_4 + FIL_1 @@ -1113,40 +922,24 @@ - + - - - - - - - - - true - + - BLOCK__OFFSET__INSIDE__FILE + key s - - - bigint - - + - - - true - + - INPUT__FILE__NAME + value s @@ -1206,7 +999,50 @@ - + + + + + + + + + + + true + + + BLOCK__OFFSET__INSIDE__FILE + + + s + + + + + bigint + + + + + + + + + true + + + INPUT__FILE__NAME + + + s + + + + + + + @@ -1220,7 +1056,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket s @@ -1232,10 +1068,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket - srcbucket0.txt + srcbucket org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1293,11 +1129,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011867 + 1290531187 Index: ql/src/test/results/compiler/plan/sample3.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample3.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample3.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011890 + 1290531191 @@ -159,21 +159,21 @@ - CNTR_NAME_FS_6_NUM_INPUT_ROWS + CNTR_NAME_FS_5_NUM_INPUT_ROWS - CNTR_NAME_FS_6_NUM_OUTPUT_ROWS + CNTR_NAME_FS_5_NUM_OUTPUT_ROWS - CNTR_NAME_FS_6_TIME_TAKEN + CNTR_NAME_FS_5_TIME_TAKEN - CNTR_NAME_FS_6_FATAL_ERROR + CNTR_NAME_FS_5_FATAL_ERROR - FS_6 + FS_5 @@ -227,21 +227,21 @@ - CNTR_NAME_TS_5_NUM_INPUT_ROWS + CNTR_NAME_TS_4_NUM_INPUT_ROWS - CNTR_NAME_TS_5_NUM_OUTPUT_ROWS + CNTR_NAME_TS_4_NUM_OUTPUT_ROWS - CNTR_NAME_TS_5_TIME_TAKEN + CNTR_NAME_TS_4_TIME_TAKEN - CNTR_NAME_TS_5_FATAL_ERROR + CNTR_NAME_TS_4_FATAL_ERROR - TS_5 + TS_4 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011890 + 1290531191 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 @@ -525,11 +525,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011883 + 1290531190 @@ -570,336 +570,98 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-11_316_6645928528775937351/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - s - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10002 - - _col0 - - - key - - - s - - - - - int - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-11_826_4879837239109698883/-ext-10000/ - - true + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - s - - - - - - - - - - _col1 - - - s - - - - - - - - - + - - - - true + + + + _col1 + + + value + + + s + + + + + - - - - - - - - - - - - - - - - - - - - key - - - s - - - - - - - - - - value - - - s - - - - - - - - - - - - - - - - - - - - - - - 2147483647 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPBitAnd - - - & - - - - - - - - - - - - - - - 2 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPMod - - - % - - - - - - - - - - - - - - - 0 - - - - + + _col0 + + + key - - + + s - + - boolean + int @@ -907,24 +669,51 @@ + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -938,9 +727,9 @@ - + - key + _col0 s @@ -951,9 +740,9 @@ - + - value + _col1 s @@ -973,6 +762,9 @@ + + true + @@ -1101,7 +893,11 @@ - + + + boolean + + @@ -1110,21 +906,21 @@ - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_FIL_1_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_FIL_1_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_FIL_1_FATAL_ERROR - FIL_4 + FIL_1 @@ -1136,40 +932,24 @@ - + - - - - - - - - - true - + - BLOCK__OFFSET__INSIDE__FILE + key s - - - bigint - - + - - - true - + - INPUT__FILE__NAME + value s @@ -1229,7 +1009,50 @@ - + + + + + + + + + + + true + + + BLOCK__OFFSET__INSIDE__FILE + + + s + + + + + bigint + + + + + + + + + true + + + INPUT__FILE__NAME + + + s + + + + + + + @@ -1243,7 +1066,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket s @@ -1255,7 +1078,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket srcbucket @@ -1316,11 +1139,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011883 + 1290531190 Index: ql/src/test/results/compiler/plan/sample4.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample4.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample4.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011906 + 1290531194 @@ -159,21 +159,21 @@ - CNTR_NAME_FS_6_NUM_INPUT_ROWS + CNTR_NAME_FS_5_NUM_INPUT_ROWS - CNTR_NAME_FS_6_NUM_OUTPUT_ROWS + CNTR_NAME_FS_5_NUM_OUTPUT_ROWS - CNTR_NAME_FS_6_TIME_TAKEN + CNTR_NAME_FS_5_TIME_TAKEN - CNTR_NAME_FS_6_FATAL_ERROR + CNTR_NAME_FS_5_FATAL_ERROR - FS_6 + FS_5 @@ -227,21 +227,21 @@ - CNTR_NAME_TS_5_NUM_INPUT_ROWS + CNTR_NAME_TS_4_NUM_INPUT_ROWS - CNTR_NAME_TS_5_NUM_OUTPUT_ROWS + CNTR_NAME_TS_4_NUM_OUTPUT_ROWS - CNTR_NAME_TS_5_TIME_TAKEN + CNTR_NAME_TS_4_TIME_TAKEN - CNTR_NAME_TS_5_FATAL_ERROR + CNTR_NAME_TS_4_FATAL_ERROR - TS_5 + TS_4 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011906 + 1290531194 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 @@ -525,11 +525,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011898 + 1290531193 @@ -570,351 +570,150 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-26_519_1779071533703199221/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - s - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10002 - - _col0 - - - key - - - s - - - - - int - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-14_730_2030189363746363960/-ext-10000/ - - true + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - s - - - - - - - - - - _col1 - - - s - - - - - - - - - + - - - - true + + + + _col1 + + + value + + + s + + + + + - - - - - - - - - - - - - - - - - - - - key - - - s - - - - - - - - - - - - - - - - - - - - - - - 2147483647 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPBitAnd - - - & - - - - - - - - - - - - - - - 2 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPMod - - - % - - - - - - - - - - - - - - - 0 - - - - + + _col0 + + + key - - + + s - + - boolean + int - - + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -928,9 +727,9 @@ - + - key + _col0 s @@ -941,9 +740,9 @@ - + - value + _col1 s @@ -963,6 +762,9 @@ + + true + @@ -1078,30 +880,37 @@ - + + + boolean + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_FIL_1_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_FIL_1_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_FIL_1_FATAL_ERROR - FIL_4 + FIL_1 @@ -1113,40 +922,24 @@ - + - - - - - - - - - true - + - BLOCK__OFFSET__INSIDE__FILE + key s - - - bigint - - + - - - true - + - INPUT__FILE__NAME + value s @@ -1206,7 +999,50 @@ - + + + + + + + + + + + true + + + BLOCK__OFFSET__INSIDE__FILE + + + s + + + + + bigint + + + + + + + + + true + + + INPUT__FILE__NAME + + + s + + + + + + + @@ -1220,7 +1056,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket s @@ -1232,10 +1068,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket - srcbucket0.txt + srcbucket org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1293,11 +1129,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011898 + 1290531193 Index: ql/src/test/results/compiler/plan/sample5.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample5.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample5.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011921 + 1290531197 @@ -159,21 +159,21 @@ - CNTR_NAME_FS_6_NUM_INPUT_ROWS + CNTR_NAME_FS_5_NUM_INPUT_ROWS - CNTR_NAME_FS_6_NUM_OUTPUT_ROWS + CNTR_NAME_FS_5_NUM_OUTPUT_ROWS - CNTR_NAME_FS_6_TIME_TAKEN + CNTR_NAME_FS_5_TIME_TAKEN - CNTR_NAME_FS_6_FATAL_ERROR + CNTR_NAME_FS_5_FATAL_ERROR - FS_6 + FS_5 @@ -227,21 +227,21 @@ - CNTR_NAME_TS_5_NUM_INPUT_ROWS + CNTR_NAME_TS_4_NUM_INPUT_ROWS - CNTR_NAME_TS_5_NUM_OUTPUT_ROWS + CNTR_NAME_TS_4_NUM_OUTPUT_ROWS - CNTR_NAME_TS_5_TIME_TAKEN + CNTR_NAME_TS_4_TIME_TAKEN - CNTR_NAME_TS_5_FATAL_ERROR + CNTR_NAME_TS_4_FATAL_ERROR - TS_5 + TS_4 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011921 + 1290531197 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 @@ -525,11 +525,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011914 + 1290531196 @@ -570,323 +570,98 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-42_112_4752933884914316684/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - s - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10002 - - _col0 - - - key - - - s - - - - - int - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-17_569_3975271910328380239/-ext-10000/ - - true + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - s - - - - - - - - - - _col1 - - - s - - - - - - - - - + - - - - true + + + + _col1 + + + value + + + s + + + + + - - - - - - - - - - - - - - - - - - - - key - - - s - - - - - - - - - - - - - - - - - - - - - - - 2147483647 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPBitAnd - - - & - - - - - - - - - - - - - - - 5 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPMod - - - % - - - - - - - - - - - - - - - 0 - - - - + + _col0 + + + key - - + + s - + - boolean + int @@ -894,24 +669,51 @@ + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -925,9 +727,9 @@ - + - key + _col0 s @@ -938,9 +740,9 @@ - + - value + _col1 s @@ -960,6 +762,9 @@ + + true + @@ -1075,7 +880,11 @@ - + + + boolean + + @@ -1084,21 +893,21 @@ - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_FIL_1_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_FIL_1_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_FIL_1_FATAL_ERROR - FIL_4 + FIL_1 @@ -1110,40 +919,24 @@ - + - - - - - - - - - true - + - BLOCK__OFFSET__INSIDE__FILE + key s - - - bigint - - + - - - true - + - INPUT__FILE__NAME + value s @@ -1203,7 +996,50 @@ - + + + + + + + + + + + true + + + BLOCK__OFFSET__INSIDE__FILE + + + s + + + + + bigint + + + + + + + + + true + + + INPUT__FILE__NAME + + + s + + + + + + + @@ -1217,7 +1053,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket s @@ -1229,7 +1065,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket srcbucket @@ -1290,11 +1126,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011914 + 1290531196 Index: ql/src/test/results/compiler/plan/sample6.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample6.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample6.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011936 + 1290531200 @@ -159,21 +159,21 @@ - CNTR_NAME_FS_6_NUM_INPUT_ROWS + CNTR_NAME_FS_5_NUM_INPUT_ROWS - CNTR_NAME_FS_6_NUM_OUTPUT_ROWS + CNTR_NAME_FS_5_NUM_OUTPUT_ROWS - CNTR_NAME_FS_6_TIME_TAKEN + CNTR_NAME_FS_5_TIME_TAKEN - CNTR_NAME_FS_6_FATAL_ERROR + CNTR_NAME_FS_5_FATAL_ERROR - FS_6 + FS_5 @@ -227,21 +227,21 @@ - CNTR_NAME_TS_5_NUM_INPUT_ROWS + CNTR_NAME_TS_4_NUM_INPUT_ROWS - CNTR_NAME_TS_5_NUM_OUTPUT_ROWS + CNTR_NAME_TS_4_NUM_OUTPUT_ROWS - CNTR_NAME_TS_5_TIME_TAKEN + CNTR_NAME_TS_4_TIME_TAKEN - CNTR_NAME_TS_5_FATAL_ERROR + CNTR_NAME_TS_4_FATAL_ERROR - TS_5 + TS_4 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011936 + 1290531200 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 @@ -525,11 +525,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011929 + 1290531199 @@ -570,351 +570,150 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-38-57_372_6472634240036667424/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - s - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10002 - - _col0 - - - key - - - s - - - - - int - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-20_565_2418859219011650919/-ext-10000/ - - true + + + + 1 + - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + - - - - - - - _col0 - - - s - - - - - - - - - - _col1 - - - s - - - - - - - - - + - - - - true + + + + _col1 + + + value + + + s + + + + + - - - - - - - - - - - - - - - - - - - - key - - - s - - - - - - - - - - - - - - - - - - - - - - - 2147483647 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPBitAnd - - - & - - - - - - - - - - - - - - - 4 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPMod - - - % - - - - - - - - - - - - - - - 0 - - - - + + _col0 + + + key - - + + s - + - boolean + int - - + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -928,9 +727,9 @@ - + - key + _col0 s @@ -941,9 +740,9 @@ - + - value + _col1 s @@ -963,6 +762,9 @@ + + true + @@ -1078,30 +880,37 @@ - + + + boolean + + + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_FIL_1_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_FIL_1_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_FIL_1_FATAL_ERROR - FIL_4 + FIL_1 @@ -1113,40 +922,24 @@ - + - - - - - - - - - true - + - BLOCK__OFFSET__INSIDE__FILE + key s - - - bigint - - + - - - true - + - INPUT__FILE__NAME + value s @@ -1206,7 +999,50 @@ - + + + + + + + + + + + true + + + BLOCK__OFFSET__INSIDE__FILE + + + s + + + + + bigint + + + + + + + + + true + + + INPUT__FILE__NAME + + + s + + + + + + + @@ -1220,7 +1056,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket s @@ -1232,10 +1068,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket - srcbucket0.txt + srcbucket org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1293,11 +1129,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011929 + 1290531199 Index: ql/src/test/results/compiler/plan/sample7.q.xml =================================================================== --- ql/src/test/results/compiler/plan/sample7.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/sample7.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -33,7 +33,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10000/ + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10000/ @@ -73,7 +73,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 @@ -82,7 +82,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10000 1 @@ -134,7 +134,7 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 file.outputformat @@ -142,7 +142,7 @@ transient_lastDdlTime - 1290011952 + 1290531203 @@ -259,10 +259,10 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 @@ -271,7 +271,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 -ext-10002 @@ -325,11 +325,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/dest1 + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/dest1 transient_lastDdlTime - 1290011952 + 1290531203 @@ -379,13 +379,13 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10000 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10001 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10001 @@ -406,10 +406,10 @@ true - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10000 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10000 @@ -435,7 +435,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 @@ -525,11 +525,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011944 + 1290531202 @@ -574,238 +574,98 @@ - + - - - - - - - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10002 - - - true - - - 1 - - - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-12_819_1412266764150682086/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_4_NUM_INPUT_ROWS - - - CNTR_NAME_FS_4_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_4_TIME_TAKEN - - - CNTR_NAME_FS_4_FATAL_ERROR - - - - - FS_4 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - value - - - s - - - - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10002 - - _col0 - - - key - - - s - - - - - int - - - - + + true - - - - - - - - - - - - - + + 1 - - - - _col0 - - - _col1 - - + + pfile:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-23_462_4369801294024155264/-ext-10000/ - - true + + + + 1 + - CNTR_NAME_SEL_3_NUM_INPUT_ROWS + CNTR_NAME_FS_4_NUM_INPUT_ROWS - CNTR_NAME_SEL_3_NUM_OUTPUT_ROWS + CNTR_NAME_FS_4_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_3_TIME_TAKEN + CNTR_NAME_FS_4_TIME_TAKEN - CNTR_NAME_SEL_3_FATAL_ERROR + CNTR_NAME_FS_4_FATAL_ERROR - SEL_3 + FS_4 - + - - - - - - - _col0 - - - s - - - - - - - - - - _col1 - - - s - - - - - - - - - + - - - - - - - - - - key - - - s - - - - - - - - - - - - - 100 - - - - + + + + _col1 + + + value - - + + s - + + + + + + _col0 + + + key + + + s + + + - boolean + int @@ -813,24 +673,51 @@ + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + + + - CNTR_NAME_FIL_2_NUM_INPUT_ROWS + CNTR_NAME_SEL_3_NUM_INPUT_ROWS - CNTR_NAME_FIL_2_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_3_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_2_TIME_TAKEN + CNTR_NAME_SEL_3_TIME_TAKEN - CNTR_NAME_FIL_2_FATAL_ERROR + CNTR_NAME_SEL_3_FATAL_ERROR - FIL_2 + SEL_3 @@ -844,9 +731,9 @@ - + - key + _col0 s @@ -857,9 +744,9 @@ - + - value + _col1 s @@ -997,7 +884,11 @@ - + + + boolean + + @@ -1037,10 +928,30 @@ - + + + key + + + s + + + + + - + + + value + + + s + + + + + @@ -1057,165 +968,32 @@ - - - - - - - - - - - - - - - - - - - key - - - s - - - - - - - - - - - - - - - - - - - - - - - 2147483647 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPBitAnd - - - & - - - - - - - - - - - - - - - 4 - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFOPMod - - - % - - - - - - - - - - - - - - - 0 - - - - + + + key - - + + s - + - - - - - - - key - - - s - - - - - - - - - - - - - 100 - - - - - - - - + - + + + 100 + - + @@ -1360,7 +1138,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt s @@ -1372,7 +1150,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket/srcbucket0.txt srcbucket0.txt @@ -1433,11 +1211,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/srcbucket transient_lastDdlTime - 1290011944 + 1290531202 Index: ql/src/test/results/compiler/plan/subq.q.xml =================================================================== --- ql/src/test/results/compiler/plan/subq.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/subq.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -41,7 +41,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 @@ -50,7 +50,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10000 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10000 1 @@ -192,10 +192,10 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 @@ -204,7 +204,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 -ext-10001 @@ -277,7 +277,7 @@ true - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10000 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10000 ../build/ql/test/data/warehouse/union.out @@ -301,10 +301,10 @@ true - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10000 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10000 @@ -330,7 +330,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 @@ -416,11 +416,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011964 + 1290531205 @@ -461,198 +461,65 @@ - + - + - - - - - - - - - 1 - - - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10001 - - - 1 - - - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-39-28_369_4065673259860755711/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_4_NUM_INPUT_ROWS - - - CNTR_NAME_FS_4_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_4_TIME_TAKEN - - - CNTR_NAME_FS_4_FATAL_ERROR - - - - - FS_4 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - _col1 - - - src - - - - - + + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10001 - - _col0 - - - _col0 - - - src - - - - - + + 1 - - - - - - - - - - - - - + + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-26_494_2054996586290574419/-ext-10000/ - - - - _col0 - - - _col1 - - + + - - true + + 1 - CNTR_NAME_SEL_3_NUM_INPUT_ROWS + CNTR_NAME_FS_4_NUM_INPUT_ROWS - CNTR_NAME_SEL_3_NUM_OUTPUT_ROWS + CNTR_NAME_FS_4_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_3_TIME_TAKEN + CNTR_NAME_FS_4_TIME_TAKEN - CNTR_NAME_SEL_3_FATAL_ERROR + CNTR_NAME_FS_4_FATAL_ERROR - SEL_3 + FS_4 - + - - - - - - - key - - - _col0 - - - src - - - - - - - - - - value - - - _col1 - - - src - - - - - - - - - + @@ -662,9 +529,9 @@ _col1 - + - value + _col1 src @@ -676,9 +543,9 @@ _col0 - + - key + _col0 src @@ -695,10 +562,10 @@ - + - + @@ -720,26 +587,26 @@ - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_SEL_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_SEL_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_SEL_3_FATAL_ERROR - SEL_2 + SEL_3 - + @@ -749,6 +616,9 @@ + + key + _col0 @@ -762,6 +632,9 @@ + + value + _col1 @@ -781,73 +654,83 @@ - - - - - - - - - - key - - - src - - - - - - - - - - - - int - - - - - 100 - - - - + + + + _col1 + + + value - - + + src - - - boolean - - + + + _col0 + + + key + + + src + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -861,9 +744,9 @@ - + - key + _col0 src @@ -874,9 +757,9 @@ - + - value + _col1 src @@ -916,7 +799,11 @@ - + + + int + + 100 @@ -929,7 +816,11 @@ - + + + boolean + + @@ -966,10 +857,30 @@ - + + + key + + + src + + + + + - + + + value + + + src + + + + + @@ -1068,7 +979,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src unioninput:src @@ -1080,7 +991,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1137,11 +1048,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011964 + 1290531205 Index: ql/src/test/results/compiler/plan/udf1.q.xml =================================================================== --- ql/src/test/results/compiler/plan/udf1.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/udf1.q.xml (working copy) @@ -1,5 +1,5 @@ - + Stage-3 @@ -62,11 +62,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011979 + 1290531208 @@ -107,1261 +107,79 @@ - + - - - - - - - - - file:/tmp/njain/hive_2010-11-17_08-39-43_567_6512311296185018388/-ext-10001 - - - 1 - - - file:/tmp/njain/hive_2010-11-17_08-39-43_567_6512311296185018388/-ext-10001/ - - - - - org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - org.apache.hadoop.mapred.TextInputFormat - - - org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - - columns - _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16 - - - serialization.format - 1 - - - columns.types - boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:string:string:string:string - - - - - - - 1 - - - - - - - CNTR_NAME_FS_3_NUM_INPUT_ROWS - - - CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_3_TIME_TAKEN - - - CNTR_NAME_FS_3_FATAL_ERROR - - - - - FS_3 - - - - - - - - - - - - - - - - _col0 - - - - - - - - boolean - - - - - - - - - _col1 - - - - - - - - - - - - - _col2 - - - - - - - - - - - - - _col3 - - - - - - - - - - - - - _col4 - - - - - - - - - - - - - _col5 - - - - - - - - - - - - - _col6 - - - - - - - - - - - - - _col7 - - - - - - - - - - - - - _col8 - - - - - - - - - - - - - _col9 - - - - - - - - - - - - - _col10 - - - - - - - - - - - - - _col11 - - - - - - - - - - - - - _col12 - - - - - - - - - - - - - _col13 - - - - - - - - string - - - - - - - - - _col14 - - - - - - - - - - - - - _col15 - - - - - - - - - - - - - _col16 - - - - - - - - - - - - - - + + + + + file:/tmp/amarsri/hive_2010-11-23_08-53-29_374_5543020152222227939/-ext-10001 - - - - - - _col8 - - - - - - - - - - - - - - - - - - - - .* - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFRegExp - - - rlike - - - - - - - + + 1 - - _col7 - - - - - - - - - - ab - - - - - - - - - - a - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - - + + file:/tmp/amarsri/hive_2010-11-23_08-53-29_374_5543020152222227939/-ext-10001/ - - _col6 - - - - - - - - - - ab - - - - - - - - - - _a% - - - - + + + + org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - + + org.apache.hadoop.mapred.TextInputFormat - - + + org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - - - _col5 - - - - - - - - - - ab - - + + + + columns + _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16 - - - - - - - \%\_ - - + + serialization.format + 1 - - - - - - true + + columns.types + boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:boolean:string:string:string:string - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - _col4 - - - - - - - - - - %_ - - - - - - - - - - \%\_ - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - - + + 1 - - _col3 - - - - - - - - - - ab - - - - - - - - - - %a_ - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - - - - - _col2 - - - - - - - - - - ab - - - - - - - - - - %a% - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - - - - - _col1 - - - - - - - - - - b - - - - - - - - - - %a% - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - - - - - _col9 - - - - - - - - - - a - - - - - - - - - - [ab] - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFRegExp - - - rlike - - - - - - - - - - _col13 - - - - - - - - - - abc - - - - - - - - - - b - - - - - - - - - - c - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFRegExpReplace - - - regexp_replace - - - - - - - - - - _col12 - - - - - - - - - - hadoop - - - - - - - - - - o* - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFRegExp - - - rlike - - - - - - - - - - _col11 - - - - - - - - - - hadoop - - - - - - - - - - [a-z]* - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFRegExp - - - rlike - - - - - - - - - - _col10 - - - - - - - - - - - - - - - - - - - - [ab] - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFRegExp - - - rlike - - - - - - - - - - _col16 - - - - - - - - - - hadoop - - - - - - - - - - (.)[a-z]* - - - - - - - - - - $1ive - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFRegExpReplace - - - regexp_replace - - - - - - - - - - _col15 - - - - - - - - - - abbbb - - - - - - - - - - bb - - - - - - - - - - b - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFRegExpReplace - - - regexp_replace - - - - - - - - - - _col14 - - - - - - - - - - abc - - - - - - - - - - z - - - - - - - - - - a - - - - - - - - - org.apache.hadoop.hive.ql.udf.UDFRegExpReplace - - - regexp_replace - - - - - - - - - - _col0 - - - - - - - - - - a - - - - - - - - - - %a% - - - - - - - - - true - - - org.apache.hadoop.hive.ql.udf.UDFLike - - - like - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - _col0 - - - _col1 - - - _col2 - - - _col3 - - - _col4 - - - _col5 - - - _col6 - - - _col7 - - - _col8 - - - _col9 - - - _col10 - - - _col11 - - - _col12 - - - _col13 - - - _col14 - - - _col15 - - - _col16 - - - - - - CNTR_NAME_SEL_2_NUM_INPUT_ROWS + CNTR_NAME_FS_3_NUM_INPUT_ROWS - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS + CNTR_NAME_FS_3_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_2_TIME_TAKEN + CNTR_NAME_FS_3_TIME_TAKEN - CNTR_NAME_SEL_2_FATAL_ERROR + CNTR_NAME_FS_3_FATAL_ERROR - SEL_2 + FS_3 - + @@ -1371,25 +189,29 @@ - - _c0 - _col0 + + + - + + + boolean + + - - _c1 - _col1 + + + @@ -1397,12 +219,12 @@ - - _c2 - _col2 + + + @@ -1410,12 +232,12 @@ - - _c3 - _col3 + + + @@ -1423,12 +245,12 @@ - - _c4 - _col4 + + + @@ -1436,12 +258,12 @@ - - _c5 - _col5 + + + @@ -1449,12 +271,12 @@ - - _c6 - _col6 + + + @@ -1462,12 +284,12 @@ - - _c7 - _col7 + + + @@ -1475,12 +297,12 @@ - - _c8 - _col8 + + + @@ -1488,12 +310,12 @@ - - _c9 - _col9 + + + @@ -1501,12 +323,12 @@ - - _c10 - _col10 + + + @@ -1514,12 +336,12 @@ - - _c11 - _col11 + + + @@ -1527,12 +349,12 @@ - - _c12 - _col12 + + + @@ -1540,25 +362,29 @@ - - _c13 - _col13 + + + - + + + string + + - - _c14 - _col14 + + + @@ -1566,12 +392,12 @@ - - _c15 - _col15 + + + @@ -1579,12 +405,12 @@ - - _c16 - _col16 + + + @@ -1598,69 +424,935 @@ - - - - + + + + _col8 + - - - key + + + - - src + + + + + + + + .* + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFRegExp + + + rlike + + + + + + + + + + _col7 + + + - - - int - - + - 86 + ab + + + + + + + a + + + - + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + _col6 + + + + + + + + + + ab + + + + + + + + + + _a% + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + _col5 + + + + + + + + + + ab + + + + + + + + + + \%\_ + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + _col4 + + + + + + + + + + %_ + + + + + + + + + + \%\_ + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + _col3 + + + + + + + + + + ab + + + + + + + + + + %a_ + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + _col2 + + + + + + + + + + ab + + + + + + + + + + %a% + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + _col1 + + + + + + + + + + b + + + + + + + + + + %a% + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + _col9 + + + + + + + + + + a + + + + + + + + + + [ab] + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFRegExp + + + rlike + + + + + + + + + + _col13 + + + + + + + + + + abc + + + + + + + + + + b + + + + + + + + + + c + + + + + + + + + org.apache.hadoop.hive.ql.udf.UDFRegExpReplace + + + regexp_replace + + + + + + + + + + _col12 + + + + + + + + + + hadoop + + + + + + + + + + o* + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFRegExp + + + rlike + + + + + + + + + + _col11 + + + + + + + + + + hadoop + + + + + + + + + + [a-z]* + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFRegExp + + + rlike + + + + + + + + + + _col10 + + + + + + + + + + + + + + + + + + + + [ab] + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFRegExp + + + rlike + + + + + + + + + + _col16 + + + + + + + + + + hadoop + + + + + + + + + + (.)[a-z]* + + + + + + + + + + $1ive + + + + + + + + + org.apache.hadoop.hive.ql.udf.UDFRegExpReplace + + + regexp_replace + + + + + + + + + + _col15 + + + + + + + + + + abbbb + + + + + + + + + + bb + + + + + + + + + + b + + + + + + + + + org.apache.hadoop.hive.ql.udf.UDFRegExpReplace + + + regexp_replace + + + + + + + + + + _col14 + + + + + + + + + + abc + + + + + + + + + + z + + + + + + + + + + a + + + + + + + + + org.apache.hadoop.hive.ql.udf.UDFRegExpReplace + + + regexp_replace + + + + + + + + + + _col0 + + + + + + + + + + a + + + + + + + + + + %a% + + + + + + + + + true + + + org.apache.hadoop.hive.ql.udf.UDFLike + + + like + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + _col2 + + + _col3 + + + _col4 + + + _col5 + + + _col6 + + + _col7 + + + _col8 + + + _col9 + + + _col10 + + + _col11 + + + _col12 + + + _col13 + + + _col14 + + + _col15 + + + _col16 + + + + + - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -1674,18 +1366,226 @@ - + + + _c0 + - key + _col0 - - src + + + + + + + + _c1 + + + _col1 + + + + + + + + + _c2 + + + _col2 + + + + + + + + + + _c3 + + + _col3 + + + + + + + + + + _c4 + + + _col4 + + + + + + + + + + _c5 + + + _col5 + + + + + + + + + + _c6 + + + _col6 + + + + + + + + + + _c7 + + + _col7 + + + + + + + + + + _c8 + + + _col8 + + + + + + + + + + _c9 + + + _col9 + + + + + + + + + + _c10 + + + _col10 + + + + + + + + + + _c11 + + + _col11 + + + + + + + + + + _c12 + + + _col12 + + + + + + + + + + _c13 + + + _col13 + + + + + + _c14 + + + _col14 + + + + + + + + + + _c15 + + + _col15 + + + + + + + + + + _c16 + + + _col16 + + + + + + @@ -1716,7 +1616,11 @@ - + + + int + + 86 @@ -1766,7 +1670,17 @@ - + + + key + + + src + + + + + @@ -1875,7 +1789,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1887,7 +1801,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1944,11 +1858,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290011979 + 1290531208 Index: ql/src/test/results/compiler/plan/union.q.xml =================================================================== --- ql/src/test/results/compiler/plan/union.q.xml (revision 1038445) +++ ql/src/test/results/compiler/plan/union.q.xml (working copy) @@ -1,5 +1,5 @@ - + @@ -41,7 +41,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 @@ -50,7 +50,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10000 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10000 1 @@ -192,10 +192,10 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 @@ -204,7 +204,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 -ext-10001 @@ -277,7 +277,7 @@ true - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10000 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10000 ../build/ql/test/data/warehouse/union.out @@ -301,10 +301,10 @@ true - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10000 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10000 @@ -330,7 +330,7 @@ - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 @@ -416,11 +416,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290012056 + 1290531223 @@ -503,11 +503,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290012056 + 1290531223 @@ -548,620 +548,516 @@ - + - + - + - - - - - - - - - 1 - - - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10001 - - - 1 - - - file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-11-17_08-41-00_459_3600797151942406281/-ext-10000/ - - - - - - 1 - - - - - - - CNTR_NAME_FS_8_NUM_INPUT_ROWS - - - CNTR_NAME_FS_8_NUM_OUTPUT_ROWS - - - CNTR_NAME_FS_8_TIME_TAKEN - - - CNTR_NAME_FS_8_FATAL_ERROR - - - - - FS_8 - - - - - - - - - - - - + + + + + 1 - - - - - - _col1 - - - _col1 - - - src - - - - - + + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10001 - - _col0 - - - _col0 - - - src - - - - - + + 1 - - - - - - - - - - - - - + + file:/home/amarsri/workspace/hive/build/ql/scratchdir/hive_2010-11-23_08-53-43_694_1594763449287632427/-ext-10000/ - - - - _col0 - - - _col1 - - + + - - true + + 1 - CNTR_NAME_SEL_7_NUM_INPUT_ROWS + CNTR_NAME_FS_8_NUM_INPUT_ROWS - CNTR_NAME_SEL_7_NUM_OUTPUT_ROWS + CNTR_NAME_FS_8_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_7_TIME_TAKEN + CNTR_NAME_FS_8_TIME_TAKEN - CNTR_NAME_SEL_7_FATAL_ERROR + CNTR_NAME_FS_8_FATAL_ERROR - SEL_7 + FS_8 - + - - - - - - - key - - - _col0 - - - src - - - - - - - - - - value - - - _col1 - - - src - - - - - - - - - + + + + + _col1 + + + _col1 + + + src + + + + + + + + _col0 + + + _col0 + + + src + + + + + + + + - + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + + - CNTR_NAME_UNION_6_NUM_INPUT_ROWS + CNTR_NAME_SEL_7_NUM_INPUT_ROWS - CNTR_NAME_UNION_6_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_7_NUM_OUTPUT_ROWS - CNTR_NAME_UNION_6_TIME_TAKEN + CNTR_NAME_SEL_7_TIME_TAKEN - CNTR_NAME_UNION_6_FATAL_ERROR + CNTR_NAME_SEL_7_FATAL_ERROR - UNION_6 + SEL_7 - + - - - - - - + + + + + + + + + + key + + _col0 + + + src + + + + - - - + + + + value + + _col1 - - - value - - - src - - - - - - - _col0 - - - key - - - src - - - - - + + src + + + + + + + + + + + + + + + + + + CNTR_NAME_UNION_6_NUM_INPUT_ROWS + + + CNTR_NAME_UNION_6_NUM_OUTPUT_ROWS + + + CNTR_NAME_UNION_6_TIME_TAKEN + + + CNTR_NAME_UNION_6_FATAL_ERROR + + + + + UNION_6 + + + + + + + + + + + + + + + + + + + _col1 + + + value + + + src + + + + + + + + _col0 + + + key + + + src + + + + + + + + + + + + + + + + + + + + + + + + _col0 + + + _col1 + + + + + true + + + + + + + CNTR_NAME_SEL_5_NUM_INPUT_ROWS + + + CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + + + CNTR_NAME_SEL_5_TIME_TAKEN + + + CNTR_NAME_SEL_5_FATAL_ERROR + + + + + SEL_5 + + + + + + + + + + + + - - - - - + + + + + + + + + key + + + src + + + + + + + + + + + + int + + + + + 100 + + + + - - + + - - - - - - _col0 + + + + boolean + + - - _col1 - - - true - - CNTR_NAME_SEL_5_NUM_INPUT_ROWS + CNTR_NAME_FIL_10_NUM_INPUT_ROWS - CNTR_NAME_SEL_5_NUM_OUTPUT_ROWS + CNTR_NAME_FIL_10_NUM_OUTPUT_ROWS - CNTR_NAME_SEL_5_TIME_TAKEN + CNTR_NAME_FIL_10_TIME_TAKEN - CNTR_NAME_SEL_5_FATAL_ERROR + CNTR_NAME_FIL_10_FATAL_ERROR - SEL_5 + FIL_10 - + - + - - - - - - - - - key - - - src - - - - - - - - - - - - int - - - - - 100 - - - - - - - - - - - - boolean - - - - + + + src + + + - CNTR_NAME_FIL_4_NUM_INPUT_ROWS + CNTR_NAME_TS_3_NUM_INPUT_ROWS - CNTR_NAME_FIL_4_NUM_OUTPUT_ROWS + CNTR_NAME_TS_3_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_4_TIME_TAKEN + CNTR_NAME_TS_3_TIME_TAKEN - CNTR_NAME_FIL_4_FATAL_ERROR + CNTR_NAME_TS_3_FATAL_ERROR - - FIL_4 - - + - - - - - + 0 + + + 1 + + + + + TS_3 + + + + + + + + + key - - - - - - - - - - - - key - - - src - - - - - - - - - - - - - 100 - - - - - - - - - - - - + + src + + + - - - - CNTR_NAME_FIL_10_NUM_INPUT_ROWS + + + + value - - CNTR_NAME_FIL_10_NUM_OUTPUT_ROWS + + src - - CNTR_NAME_FIL_10_TIME_TAKEN + + - - CNTR_NAME_FIL_10_FATAL_ERROR - - - FIL_10 - - - - - - - - - - - + + + + true + + + BLOCK__OFFSET__INSIDE__FILE + + + src + + + + + bigint - - - - src - - - - - - - - - - CNTR_NAME_TS_3_NUM_INPUT_ROWS - - - CNTR_NAME_TS_3_NUM_OUTPUT_ROWS - - - CNTR_NAME_TS_3_TIME_TAKEN - - - CNTR_NAME_TS_3_FATAL_ERROR - - - - - - - 0 - - - 1 - - - - - TS_3 - - - - - - - - - key - - - src - - - - - - - - - - value - - - src - - - - - - - - - - true - - - BLOCK__OFFSET__INSIDE__FILE - - - src - - - - - bigint - - - - - - - - - true - - - INPUT__FILE__NAME - - - src - - - - - - - - - - - - - - + + + + true + + INPUT__FILE__NAME + + + src + + + + - - - - - - - - - - - - - - @@ -1169,34 +1065,7 @@ - - - - - _col0 - - - src - - - - - - - - - - _col1 - - - src - - - - - - - + @@ -1209,7 +1078,7 @@ - + _col0 @@ -1222,7 +1091,7 @@ - + _col1 @@ -1242,143 +1111,96 @@ - - - - _col1 - - - value - - - src - - - - - - - - _col0 - - - key - - - src - - - - - - - - - - - + + + - + + + _col0 + + + src + + + + + - + + + _col1 + + + src + + + + + - - - - _col0 - - - _col1 - - - - - - - CNTR_NAME_SEL_2_NUM_INPUT_ROWS - - - CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - - - CNTR_NAME_SEL_2_TIME_TAKEN - - - CNTR_NAME_SEL_2_FATAL_ERROR - - + + + + + + + + _col1 + + + value - - SEL_2 + + src - - - - - - + + - - - - - - - - - - - - - + + + + _col0 + + + key + + src + + + + - - - - - - - - - key - - - src - - - - - - - - - - - - - 100 - - - - + + + + + - - + + - - + + + + + + _col0 + + _col1 + @@ -1386,21 +1208,21 @@ - CNTR_NAME_FIL_1_NUM_INPUT_ROWS + CNTR_NAME_SEL_2_NUM_INPUT_ROWS - CNTR_NAME_FIL_1_NUM_OUTPUT_ROWS + CNTR_NAME_SEL_2_NUM_OUTPUT_ROWS - CNTR_NAME_FIL_1_TIME_TAKEN + CNTR_NAME_SEL_2_TIME_TAKEN - CNTR_NAME_FIL_1_FATAL_ERROR + CNTR_NAME_SEL_2_FATAL_ERROR - FIL_1 + SEL_2 @@ -1414,30 +1236,10 @@ - - - key - - - src - - - - - + - - - value - - - src - - - - - + @@ -1519,10 +1321,30 @@ - + + + key + + + src + + + + + - + + + value + + + src + + + + + @@ -1621,7 +1443,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src null-subquery1:unioninput-subquery1:src @@ -1636,7 +1458,7 @@ - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src src @@ -1693,11 +1515,11 @@ location - pfile:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/src + pfile:/home/amarsri/workspace/hive/build/ql/test/data/warehouse/src transient_lastDdlTime - 1290012056 + 1290531223