diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java index 3a51d9795b..833757c4eb 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java @@ -1636,7 +1636,7 @@ public String toString() { } } - public class AnalyzeRewriteContext { + public static class AnalyzeRewriteContext { private String tableName; private List colName; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsAutoGatherContext.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsAutoGatherContext.java index 11ccff4458..d068b0e119 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsAutoGatherContext.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsAutoGatherContext.java @@ -25,6 +25,7 @@ import java.util.Map; import org.apache.hadoop.hive.conf.HiveConf; +import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.ql.Context; import org.apache.hadoop.hive.ql.QueryState; @@ -32,6 +33,8 @@ import org.apache.hadoop.hive.ql.exec.Operator; import org.apache.hadoop.hive.ql.exec.RowSchema; import org.apache.hadoop.hive.ql.exec.SelectOperator; +import org.apache.hadoop.hive.ql.exec.UDTFOperator; +import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.AnalyzeRewriteContext; @@ -93,7 +96,7 @@ public void setAnalyzeRewrite(AnalyzeRewriteContext analyzeRewrite) { } public void insertAnalyzePipeline() throws SemanticException{ - // 1. Generate the statement of analyze table [tablename] compute statistics for columns + // Generate the statement of analyze table [tablename] compute statistics for columns // In non-partitioned table case, it will generate TS-SEL-GBY-RS-GBY-SEL-FS operator // In static-partitioned table case, it will generate TS-FIL(partitionKey)-SEL-GBY(partitionKey)-RS-GBY-SEL-FS operator // In dynamic-partitioned table case, it will generate TS-SEL-GBY(partitionKey)-RS-GBY-SEL-FS operator @@ -101,70 +104,114 @@ public void insertAnalyzePipeline() throws SemanticException{ // (2) we can compose the static/dynamic partition using a select operator in replaceSelectOperatorProcess.. String analyzeCommand = "analyze table `" + tbl.getDbName() + "`.`" + tbl.getTableName() + "`" + " compute statistics for columns "; + insertAnalyzePipeline(analyzeCommand, false); + } + + public void insertTableValuesAnalyzePipeline() throws SemanticException { + // Instead of starting from analyze statement, we just generate the Select plan + boolean isPartitionStats = conf.getBoolVar(ConfVars.HIVE_STATS_COLLECT_PART_LEVEL_STATS) && tbl.isPartitioned(); + if (isPartitionStats) { + partSpec = new HashMap<>(); + List partKeys = Utilities.getColumnNamesFromFieldSchema(tbl.getPartitionKeys()); + for (String partKey : partKeys) { + partSpec.put(partKey, null); + } + } + String command = ColumnStatsSemanticAnalyzer.genRewrittenQuery( + tbl, Utilities.getColumnNamesFromFieldSchema(tbl.getCols()), conf, partSpec, isPartitionStats, true); + insertAnalyzePipeline(command, true); + } - // 2. Based on the statement, generate the selectOperator + private void insertAnalyzePipeline(String command, boolean rewritten) throws SemanticException { + // 1. Based on the statement, generate the selectOperator Operator selOp = null; try { - selOp = genSelOpForAnalyze(analyzeCommand, origCtx); + selOp = genSelOp(command, rewritten, origCtx); } catch (IOException | ParseException e) { throw new SemanticException(e); } - // 3. attach this SEL to the operator right before FS + // 2. attach this SEL to the operator right before FS op.getChildOperators().add(selOp); selOp.getParentOperators().clear(); selOp.getParentOperators().add(op); - // 4. address the colExp, colList, etc for the SEL + // 3. address the colExp, colList, etc for the SEL try { replaceSelectOperatorProcess((SelectOperator)selOp, op); } catch (HiveException e) { throw new SemanticException(e); } } - - @SuppressWarnings("rawtypes") - private Operator genSelOpForAnalyze(String analyzeCommand, Context origCtx) throws IOException, ParseException, SemanticException{ - //0. initialization + + private Operator genSelOp(String command, boolean rewritten, Context origCtx) + throws IOException, ParseException, SemanticException { + // 1. initialization Context ctx = new Context(conf); ctx.setOpContext(origCtx.getOpContext()); ctx.setExplainConfig(origCtx.getExplainConfig()); - ASTNode tree = ParseUtils.parse(analyzeCommand, ctx); - //1. get the ColumnStatsSemanticAnalyzer - QueryState queryState = new QueryState.Builder().withHiveConf(conf).build(); - BaseSemanticAnalyzer baseSem = SemanticAnalyzerFactory.get(queryState, tree); - ColumnStatsSemanticAnalyzer colSem = (ColumnStatsSemanticAnalyzer) baseSem; + // 2. parse tree and create semantic analyzer. if we need to rewrite the analyze + // statement, we do it now + final ASTNode ast; + final SemanticAnalyzer sem; + final QueryState queryState = new QueryState.Builder().withHiveConf(conf).build(); + if (rewritten) { + // Create the context object that is needed to store the column stats + this.analyzeRewrite = ColumnStatsSemanticAnalyzer.genAnalyzeRewriteContext(conf, tbl); + + // The analyze statement has already been rewritten, we just need to create the AST + // and the corresponding semantic analyzer + ast = ParseUtils.parse(command, ctx); + BaseSemanticAnalyzer baseSem = SemanticAnalyzerFactory.get(queryState, ast); + sem = (SemanticAnalyzer) baseSem; + } else { + // We need to rewrite the analyze command and get the rewritten AST + ASTNode analyzeTree = ParseUtils.parse(command, ctx); + BaseSemanticAnalyzer baseSem = SemanticAnalyzerFactory.get(queryState, analyzeTree); + ColumnStatsSemanticAnalyzer colSem = (ColumnStatsSemanticAnalyzer) baseSem; + ast = colSem.rewriteAST(analyzeTree, this); + + // Obtain the context object that is needed to store the column stats + this.analyzeRewrite = colSem.getAnalyzeRewriteContext(); - //2. get the rewritten AST - ASTNode ast = colSem.rewriteAST(tree, this); - baseSem = SemanticAnalyzerFactory.get(queryState, ast); - SemanticAnalyzer sem = (SemanticAnalyzer) baseSem; + // Analyze the rewritten statement + baseSem = SemanticAnalyzerFactory.get(queryState, ast); + sem = (SemanticAnalyzer) baseSem; + } QB qb = new QB(null, null, false); ASTNode child = ast; - ParseContext subPCtx = ((SemanticAnalyzer) sem).getParseContext(); + ParseContext subPCtx = sem.getParseContext(); subPCtx.setContext(ctx); - ((SemanticAnalyzer) sem).initParseCtx(subPCtx); + sem.initParseCtx(subPCtx); sem.doPhase1(child, qb, sem.initPhase1Ctx(), null); // This will trigger new calls to metastore to collect metadata // TODO: cache the information from the metastore sem.getMetaData(qb); - Operator operator = sem.genPlan(qb); + sem.genPlan(qb); - //3. populate the load file work so that ColumnStatsTask can work + // 3. populate the load file work so that ColumnStatsTask can work loadFileWork.addAll(sem.getLoadFileWork()); - //4. because there is only one TS for analyze statement, we can get it. + // 4. because there is only one TS for analyze statement, we can get it. if (sem.topOps.values().size() != 1) { throw new SemanticException( "ColumnStatsAutoGatherContext is expecting exactly one TS, but finds " + sem.topOps.values().size()); } - operator = sem.topOps.values().iterator().next(); + Operator operator = sem.topOps.values().iterator().next(); - //5. get the first SEL after TS - while(!(operator instanceof SelectOperator)){ + // 5. if this has been rewritten, get the SEL after UDTF; + // otherwise, get the first SEL after TS + if (rewritten) { + while (!(operator instanceof UDTFOperator)) { + operator = operator.getChildOperators().get(0); + } operator = operator.getChildOperators().get(0); + } else { + while (!(operator instanceof SelectOperator)) { + operator = operator.getChildOperators().get(0); + } } return operator; } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java index 9aff0069fd..23f5a3388b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java @@ -148,7 +148,8 @@ private void handlePartialPartitionSpec(Map partSpec, ColumnStat } } - private StringBuilder genPartitionClause(Map partSpec) throws SemanticException { + private static StringBuilder genPartitionClause(Table tbl, Map partSpec) + throws SemanticException { StringBuilder whereClause = new StringBuilder(" where "); boolean predPresent = false; StringBuilder groupByClause = new StringBuilder(" group by "); @@ -162,7 +163,8 @@ private StringBuilder genPartitionClause(Map partSpec) throws Se } else { whereClause.append(" and "); } - whereClause.append("`").append(partKey).append("` = ").append(genPartValueString(getColTypeOf(partKey), value)); + whereClause.append("`").append(partKey).append("` = ") + .append(genPartValueString(getColTypeOf(tbl, partKey), value)); } } @@ -181,8 +183,7 @@ private StringBuilder genPartitionClause(Map partSpec) throws Se - private String getColTypeOf(String partKey) throws SemanticException{ - + private static String getColTypeOf(Table tbl, String partKey) throws SemanticException{ for (FieldSchema fs : tbl.getPartitionKeys()) { if (partKey.equalsIgnoreCase(fs.getName())) { return fs.getType().toLowerCase(); @@ -191,8 +192,7 @@ private String getColTypeOf(String partKey) throws SemanticException{ throw new SemanticException("Unknown partition key : " + partKey); } - private List getColumnTypes(List colNames) - throws SemanticException{ + private static List getColumnTypes(Table tbl, List colNames) { List colTypes = new ArrayList(); List cols = tbl.getCols(); List copyColNames = new ArrayList<>(colNames); @@ -215,21 +215,33 @@ private String getColTypeOf(String partKey) throws SemanticException{ return colTypes; } - private String escapeBackTicks(String colName) { + private static String escapeBackTicks(String colName) { return colName.replaceAll("`", "``"); } private String genRewrittenQuery(List colNames, HiveConf conf, Map partSpec, - boolean isPartitionStats) throws SemanticException{ + boolean isPartitionStats, boolean useTableValues) throws SemanticException { + String rewrittenQuery = genRewrittenQuery(tbl, colNames, conf, partSpec, isPartitionStats, useTableValues); + isRewritten = true; + return rewrittenQuery; + } + + public static String genRewrittenQuery(Table tbl, List colNames, HiveConf conf, Map partSpec, + boolean isPartitionStats, boolean useTableValues) throws SemanticException{ StringBuilder rewrittenQueryBuilder = new StringBuilder("select "); + StringBuilder columnNamesBuilder = new StringBuilder(); + StringBuilder columnDummyValuesBuilder = new StringBuilder(); for (int i = 0; i < colNames.size(); i++) { if (i > 0) { rewrittenQueryBuilder.append(" , "); + columnNamesBuilder.append(" , "); + columnDummyValuesBuilder.append(" , "); } String func = HiveConf.getVar(conf, HiveConf.ConfVars.HIVE_STATS_NDV_ALGO).toLowerCase(); rewrittenQueryBuilder.append("compute_stats(`"); - rewrittenQueryBuilder.append(escapeBackTicks(colNames.get(i))); + final String columnName = escapeBackTicks(colNames.get(i)); + rewrittenQueryBuilder.append(columnName); rewrittenQueryBuilder.append("`, '" + func + "'"); if ("fm".equals(func)) { int numBitVectors = 0; @@ -241,23 +253,45 @@ private String genRewrittenQuery(List colNames, HiveConf conf, Map specifiedCols, List p } } - private void logTypeWarning(String colName, String colType) { + private static void logTypeWarning(String colName, String colType) { String warning = "Only primitive type arguments are accepted but " + colType + " is passed for " + colName + "."; warning = "WARNING: " + warning; @@ -353,8 +387,8 @@ public void analyze(ASTNode ast, Context origCtx) throws SemanticException { } else { isTableLevel = true; } - colType = getColumnTypes(colNames); - rewrittenQuery = genRewrittenQuery(colNames, conf, partSpec, isPartitionStats); + colType = getColumnTypes(tbl, colNames); + rewrittenQuery = genRewrittenQuery(colNames, conf, partSpec, isPartitionStats, false); rewrittenTree = genRewrittenTree(rewrittenQuery); } else { // Not an analyze table column compute statistics statement - don't do any rewrites @@ -391,19 +425,19 @@ public void analyze(ASTNode ast, Context origCtx) throws SemanticException { /** * @param ast * is the original analyze ast - * @param qb - * is the qb that calls this function - * @param sem - * is the semantic analyzer that calls this function + * @param context + * the column stats auto gather context * @return * @throws SemanticException */ public ASTNode rewriteAST(ASTNode ast, ColumnStatsAutoGatherContext context) throws SemanticException { - tbl = AnalyzeCommandUtils.getTable(ast, this); - colNames = getColumnName(ast); // Save away the original AST originalTree = ast; + + tbl = AnalyzeCommandUtils.getTable(ast, this); + + colNames = getColumnName(ast); boolean isPartitionStats = AnalyzeCommandUtils.isPartitionLevelStats(ast); Map partSpec = null; checkForPartitionColumns(colNames, @@ -414,21 +448,38 @@ public ASTNode rewriteAST(ASTNode ast, ColumnStatsAutoGatherContext context) } if (isPartitionStats) { - isTableLevel = false; partSpec = AnalyzeCommandUtils.getPartKeyValuePairsFromAST(tbl, ast, conf); handlePartialPartitionSpec(partSpec, context); - } else { - isTableLevel = true; } - colType = getColumnTypes(colNames); - rewrittenQuery = genRewrittenQuery(colNames, conf, partSpec, isPartitionStats); + + colType = getColumnTypes(tbl, colNames); + + isTableLevel = !isPartitionStats; + + rewrittenQuery = genRewrittenQuery(colNames, conf, partSpec, isPartitionStats, false); rewrittenTree = genRewrittenTree(rewrittenQuery); - context.analyzeRewrite = new AnalyzeRewriteContext(); - context.analyzeRewrite.setTableName(tbl.getFullyQualifiedName()); - context.analyzeRewrite.setTblLvl(isTableLevel); - context.analyzeRewrite.setColName(colNames); - context.analyzeRewrite.setColType(colType); return rewrittenTree; } + + AnalyzeRewriteContext getAnalyzeRewriteContext() { + AnalyzeRewriteContext analyzeRewrite = new AnalyzeRewriteContext(); + analyzeRewrite.setTableName(tbl.getFullyQualifiedName()); + analyzeRewrite.setTblLvl(isTableLevel); + analyzeRewrite.setColName(colNames); + analyzeRewrite.setColType(colType); + return analyzeRewrite; + } + + static AnalyzeRewriteContext genAnalyzeRewriteContext(HiveConf conf, Table tbl) { + AnalyzeRewriteContext analyzeRewrite = new AnalyzeRewriteContext(); + analyzeRewrite.setTableName(tbl.getFullyQualifiedName()); + analyzeRewrite.setTblLvl(!(conf.getBoolVar(ConfVars.HIVE_STATS_COLLECT_PART_LEVEL_STATS) && tbl.isPartitioned())); + List colNames = Utilities.getColumnNamesFromFieldSchema(tbl.getCols()); + List colTypes = getColumnTypes(tbl, colNames); + analyzeRewrite.setColName(colNames); + analyzeRewrite.setColType(colTypes); + return analyzeRewrite; + } + } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index b330d710a1..d1c9c12a2b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -543,7 +543,7 @@ public CompilationOpContext getOpContext() { return ctx.getOpContext(); } - public String genPartValueString(String partColType, String partVal) throws SemanticException { + public static String genPartValueString(String partColType, String partVal) throws SemanticException { String returnVal = partVal; if (partColType.equals(serdeConstants.STRING_TYPE_NAME) || partColType.contains(serdeConstants.VARCHAR_TYPE_NAME) || @@ -7495,9 +7495,6 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) viewDesc.setPartCols(new ArrayList<>(partitionColumns)); } - destTableIsTransactional = tblDesc != null && AcidUtils.isTransactionalTable(tblDesc); - destTableIsFullAcid = tblDesc != null && AcidUtils.isFullAcidTable(tblDesc); - boolean isDestTempFile = true; if (!ctx.isMRTmpFileURI(destinationPath.toUri().toString())) { idToTableNameMap.put(String.valueOf(destTableId), destinationPath.toUri().toString()); @@ -7542,7 +7539,15 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) tableDescriptor = PlanUtils.getTableDesc(tblDesc, cols, colTypes); } - boolean isDfsDir = (destType.intValue() == QBMetaData.DEST_DFS_FILE); + boolean isDfsDir = (destType == QBMetaData.DEST_DFS_FILE); + + try { + destinationTable = tblDesc != null ? tblDesc.toTable(conf) : viewDesc != null ? viewDesc.toTable(conf) : null; + } catch (HiveException e) { + throw new SemanticException(e); + } + + destTableIsFullAcid = AcidUtils.isFullAcidTable(destinationTable); if (isPartitioned) { // Create a SELECT that may reorder the columns if needed @@ -7564,12 +7569,6 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) .getColumnInfos()), input), rowResolver); input.setColumnExprMap(colExprMap); - try { - destinationTable = tblDesc != null ? tblDesc.toTable(conf) : viewDesc.toTable(conf); - } catch (HiveException e) { - throw new SemanticException(e); - } - // If this is a partitioned CTAS or MV statement, we are going to create a LoadTableDesc // object. Although the table does not exist in metastore, we will swamp the CreateTableTask // and MoveTask resulting from this LoadTable so in this specific case, first we create @@ -7622,7 +7621,7 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) throw new SemanticException("Unknown destination type: " + destType); } - if (!(destType.intValue() == QBMetaData.DEST_DFS_FILE && qb.getIsQuery())) { + if (!(destType == QBMetaData.DEST_DFS_FILE && qb.getIsQuery())) { input = genConversionSelectOperator(dest, qb, input, tableDescriptor, dpCtx); } @@ -7661,8 +7660,7 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) canBeMerged &= !destTableIsFullAcid; // Generate the partition columns from the parent input - if (destType.intValue() == QBMetaData.DEST_TABLE - || destType.intValue() == QBMetaData.DEST_PARTITION) { + if (destType == QBMetaData.DEST_TABLE || destType == QBMetaData.DEST_PARTITION) { genPartnCols(dest, input, qb, tableDescriptor, destinationTable, rsCtx); } @@ -7704,14 +7702,20 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) // and it is an insert overwrite or insert into table if (conf.getBoolVar(ConfVars.HIVESTATSAUTOGATHER) && conf.getBoolVar(ConfVars.HIVESTATSCOLAUTOGATHER) + && destinationTable != null && !destinationTable.isNonNative() && ColumnStatsAutoGatherContext.canRunAutogatherStats(fso)) { - // TODO: Column stats autogather does not work for CTAS statements - if (destType.intValue() == QBMetaData.DEST_TABLE && !destinationTable.isNonNative()) { - genAutoColumnStatsGatheringPipeline(qb, destinationTable, partSpec, input, qb.getParseInfo() - .isInsertIntoTable(destinationTable.getDbName(), destinationTable.getTableName())); - } else if (destType.intValue() == QBMetaData.DEST_PARTITION && !destinationTable.isNonNative()) { - genAutoColumnStatsGatheringPipeline(qb, destinationTable, destinationPartition.getSpec(), input, qb - .getParseInfo().isInsertIntoTable(destinationTable.getDbName(), destinationTable.getTableName())); + if (destType == QBMetaData.DEST_TABLE) { + genAutoColumnStatsGatheringPipeline(qb, destinationTable, partSpec, input, + qb.getParseInfo().isInsertIntoTable(destinationTable.getDbName(), destinationTable.getTableName()), + false); + } else if (destType == QBMetaData.DEST_PARTITION) { + genAutoColumnStatsGatheringPipeline(qb, destinationTable, destinationPartition.getSpec(), input, + qb.getParseInfo().isInsertIntoTable(destinationTable.getDbName(), destinationTable.getTableName()), + false); + } else if (destType == QBMetaData.DEST_LOCAL_FILE || destType == QBMetaData.DEST_DFS_FILE) { + // CTAS or CMV statement + genAutoColumnStatsGatheringPipeline(qb, destinationTable, null, input, + false, true); } } return output; @@ -8066,13 +8070,20 @@ private void createPreInsertDesc(Table table, boolean overwrite) { } - private void genAutoColumnStatsGatheringPipeline(QB qb, Table table, - Map partSpec, Operator curr, boolean isInsertInto) throws SemanticException { + private void genAutoColumnStatsGatheringPipeline(QB qb, Table table, Map partSpec, + Operator curr, boolean isInsertInto, boolean useTableValueConstructor) + throws SemanticException { LOG.info("Generate an operator pipeline to autogather column stats for table " + table.getTableName() + " in query " + ctx.getCmd()); ColumnStatsAutoGatherContext columnStatsAutoGatherContext = null; columnStatsAutoGatherContext = new ColumnStatsAutoGatherContext(this, conf, curr, table, partSpec, isInsertInto, ctx); - columnStatsAutoGatherContext.insertAnalyzePipeline(); + if (useTableValueConstructor) { + // Table does not exist, use table value constructor to simulate + columnStatsAutoGatherContext.insertTableValuesAnalyzePipeline(); + } else { + // Table already exists + columnStatsAutoGatherContext.insertAnalyzePipeline(); + } columnStatsAutoGatherContexts.add(columnStatsAutoGatherContext); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateViewDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateViewDesc.java index f0f7b18d19..456d00bda4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateViewDesc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/CreateViewDesc.java @@ -24,9 +24,11 @@ import java.util.Set; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.TableType; import org.apache.hadoop.hive.metastore.api.FieldSchema; +import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; import org.apache.hadoop.hive.ql.exec.DDLTask; import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.metadata.HiveException; @@ -410,6 +412,9 @@ public Table toTable(HiveConf conf) throws HiveException { } } + StatsSetupConst.setStatsStateForCreateTable(tbl.getTTable().getParameters(), null, + StatsSetupConst.FALSE); + return tbl; } } diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite.q index 9735e61598..3b18932bf5 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite.q @@ -15,8 +15,6 @@ insert into cmv_basetable_n10 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n10 compute statistics for columns; - create materialized view cmv_mat_view_n10 as select a, b, c from cmv_basetable_n10 where a = 2; diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_2.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_2.q index 3f695d1ee2..998aa49151 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_2.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_2.q @@ -13,8 +13,6 @@ insert into cmv_basetable_n9 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n9 compute statistics for columns; - create materialized view cmv_mat_view_n9 as select b from cmv_basetable_n9 where c > 10.0 group by a, b, c; @@ -57,8 +55,6 @@ insert into cmv_basetable_2_n4 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2_n4 compute statistics for columns; - create materialized view cmv_mat_view_5 as select cmv_basetable_n9.a, cmv_basetable_2_n4.c from cmv_basetable_n9 join cmv_basetable_2_n4 on (cmv_basetable_n9.a = cmv_basetable_2_n4.a) diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_3.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_3.q index eb668a90ac..61d7c115f0 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_3.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_3.q @@ -13,16 +13,12 @@ insert into cmv_basetable values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable compute statistics for columns; - create table cmv_basetable_2 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into cmv_basetable_2 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2 compute statistics for columns; - EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view AS SELECT cmv_basetable.a, cmv_basetable_2.c @@ -51,8 +47,6 @@ GROUP BY cmv_basetable.a, cmv_basetable_2.c; insert into cmv_basetable_2 values (3, 'charlie', 15.8, 1); -analyze table cmv_basetable_2 compute statistics for columns; - -- CANNOT USE THE VIEW, IT IS OUTDATED EXPLAIN SELECT cmv_basetable.a @@ -119,8 +113,6 @@ insert into cmv_irrelevant_table values (1, 'alfred', 10.30, 2), (3, 'charlie', 9.8, 1); -analyze table cmv_irrelevant_table compute statistics for columns; - -- IT CAN STILL BE USED EXPLAIN SELECT cmv_basetable.a diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_4.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_4.q index f21db8a8d8..93844cda54 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_4.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_4.q @@ -13,16 +13,12 @@ insert into cmv_basetable_n5 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n5 compute statistics for columns; - create table cmv_basetable_2_n2 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into cmv_basetable_2_n2 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2_n2 compute statistics for columns; - -- CREATE VIEW WITH REWRITE DISABLED EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view_n5 DISABLE REWRITE TBLPROPERTIES ('transactional'='true') AS @@ -37,8 +33,6 @@ CREATE MATERIALIZED VIEW cmv_mat_view_n5 DISABLE REWRITE TBLPROPERTIES ('transac WHERE cmv_basetable_2_n2.c > 10.0 GROUP BY cmv_basetable_n5.a, cmv_basetable_2_n2.c; -analyze table cmv_mat_view_n5 compute statistics for columns; - DESCRIBE FORMATTED cmv_mat_view_n5; -- CANNOT USE THE VIEW, IT IS DISABLED FOR REWRITE @@ -56,8 +50,6 @@ GROUP BY cmv_basetable_n5.a, cmv_basetable_2_n2.c; insert into cmv_basetable_2_n2 values (3, 'charlie', 15.8, 1); -analyze table cmv_basetable_2_n2 compute statistics for columns; - -- ENABLE FOR REWRITE EXPLAIN ALTER MATERIALIZED VIEW cmv_mat_view_n5 ENABLE REWRITE; diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_5.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_5.q index 3026d9093e..8fdc1da4b7 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_5.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_5.q @@ -13,28 +13,21 @@ insert into cmv_basetable_n6 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n6 compute statistics for columns; - create table cmv_basetable_2_n3 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into cmv_basetable_2_n3 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2_n3 compute statistics for columns; - CREATE MATERIALIZED VIEW cmv_mat_view_n6 TBLPROPERTIES ('transactional'='true') AS SELECT cmv_basetable_n6.a, cmv_basetable_2_n3.c FROM cmv_basetable_n6 JOIN cmv_basetable_2_n3 ON (cmv_basetable_n6.a = cmv_basetable_2_n3.a) WHERE cmv_basetable_2_n3.c > 10.0; -analyze table cmv_mat_view_n6 compute statistics for columns; insert into cmv_basetable_2_n3 values (3, 'charlie', 15.8, 1); -analyze table cmv_basetable_2_n3 compute statistics for columns; - -- CANNOT USE THE VIEW, IT IS OUTDATED EXPLAIN SELECT cmv_basetable_n6.a diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_dummy.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_dummy.q index 8c9da8ae69..697859345d 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_dummy.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_dummy.q @@ -16,8 +16,6 @@ insert into cmv_basetable_n0 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n0 compute statistics for columns; - create materialized view cmv_mat_view_n0 as select a, b, c from cmv_basetable_n0 where a = 2; diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_multi_db.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_multi_db.q index 85d926f9eb..6de65aa243 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_multi_db.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_multi_db.q @@ -17,8 +17,6 @@ insert into cmv_basetable_n7 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n7 compute statistics for columns; - create database db2; use db2; diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_rebuild_dummy.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_rebuild_dummy.q index 72e3d65117..e0dcf9aa9c 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_rebuild_dummy.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_rebuild_dummy.q @@ -13,16 +13,12 @@ insert into cmv_basetable_n1 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n1 compute statistics for columns; - create table cmv_basetable_2_n0 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into cmv_basetable_2_n0 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2_n0 compute statistics for columns; - EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view_n1 AS SELECT cmv_basetable_n1.a, cmv_basetable_2_n0.c @@ -51,8 +47,6 @@ GROUP BY cmv_basetable_n1.a, cmv_basetable_2_n0.c; insert into cmv_basetable_2_n0 values (3, 'charlie', 15.8, 1); -analyze table cmv_basetable_2_n0 compute statistics for columns; - -- CANNOT USE THE VIEW, IT IS OUTDATED EXPLAIN SELECT cmv_basetable_n1.a @@ -119,8 +113,6 @@ insert into cmv_irrelevant_table_n0 values (1, 'alfred', 10.30, 2), (3, 'charlie', 9.8, 1); -analyze table cmv_irrelevant_table_n0 compute statistics for columns; - -- IT CAN STILL BE USED EXPLAIN SELECT cmv_basetable_n1.a diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window.q index 4cdb715d28..fa07401611 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window.q @@ -12,16 +12,12 @@ insert into cmv_basetable_n3 values (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n3 compute statistics for columns; - create table cmv_basetable_2_n1 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into cmv_basetable_2_n1 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2_n1 compute statistics for columns; - -- CREATE VIEW WITH REWRITE DISABLED EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view_n3 DISABLE REWRITE TBLPROPERTIES('rewriting.time.window'='5min') AS @@ -53,8 +49,6 @@ GROUP BY cmv_basetable_n3.a, cmv_basetable_2_n1.c; insert into cmv_basetable_2_n1 values (3, 'charlie', 15.8, 1); -analyze table cmv_basetable_2_n1 compute statistics for columns; - -- ENABLE FOR REWRITE EXPLAIN ALTER MATERIALIZED VIEW cmv_mat_view_n3 ENABLE REWRITE; diff --git a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window_2.q b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window_2.q index 6873673a55..418ec45c91 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window_2.q +++ b/ql/src/test/queries/clientpositive/materialized_view_create_rewrite_time_window_2.q @@ -9,13 +9,11 @@ insert into cmv_basetable_n100 values (2, 'bonnie', 172342.2, 3), (3, 'calvin', 978.76, 3), (3, 'charlie', 9.8, 1); -analyze table cmv_basetable_n100 compute statistics for columns; create table cmv_basetable_2_n100 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into cmv_basetable_2_n100 values (1, 'alfred', 10.30, 2), (3, 'calvin', 978.76, 3); -analyze table cmv_basetable_2_n100 compute statistics for columns; -- CREATE MATERIALIZED VIEW CREATE MATERIALIZED VIEW cmv_mat_view_n300 AS @@ -29,7 +27,6 @@ DESCRIBE FORMATTED cmv_mat_view_n300; insert into cmv_basetable_2_n100 values (3, 'charlie', 15.8, 1); -analyze table cmv_basetable_2_n100 compute statistics for columns; -- OUTDATED: YES DESCRIBE FORMATTED cmv_mat_view_n300; diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_1.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_1.q index 18b9f7d418..ff7cefc1f9 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_1.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_1.q @@ -15,7 +15,6 @@ create table emps_n3 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n3 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps_n3 compute statistics for columns; create table depts_n2 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n2 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n2 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n2 compute statistics for columns; create table dependents_n2 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n2 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n2 compute statistics for columns; create table locations_n2 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n2 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n2 compute statistics for columns; alter table emps_n3 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n2 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -50,7 +46,6 @@ alter table depts_n2 add constraint fk2 foreign key (locationid) references loca -- EXAMPLE 1 create materialized view mv1_n2 as select * from emps_n3 where empid < 150; -analyze table mv1_n2 compute statistics for columns; explain select * @@ -67,7 +62,6 @@ drop materialized view mv1_n2; create materialized view mv1_n2 as select deptno, name, salary, commission from emps_n3; -analyze table mv1_n2 compute statistics for columns; explain select emps_n3.name, emps_n3.salary, emps_n3.commission @@ -84,7 +78,6 @@ drop materialized view mv1_n2; create materialized view mv1_n2 as select empid deptno from emps_n3 join depts_n2 using (deptno); -analyze table mv1_n2 compute statistics for columns; explain select empid deptno from emps_n3 @@ -98,7 +91,6 @@ drop materialized view mv1_n2; -- EXAMPLE 4 create materialized view mv1_n2 as select * from emps_n3 where empid < 200; -analyze table mv1_n2 compute statistics for columns; explain select * from emps_n3 where empid > 120 @@ -112,7 +104,6 @@ drop materialized view mv1_n2; -- EXAMPLE 5 - NO MV, ALREADY UNIQUE create materialized view mv1_n2 as select empid, deptno from emps_n3 group by empid, deptno; -analyze table mv1_n2 compute statistics for columns; explain select empid, deptno from emps_n3 group by empid, deptno; @@ -124,7 +115,6 @@ drop materialized view mv1_n2; -- EXAMPLE 5 - NO MV, ALREADY UNIQUE create materialized view mv1_n2 as select empid, name from emps_n3 group by empid, name; -analyze table mv1_n2 compute statistics for columns; explain select empid, name from emps_n3 group by empid, name; @@ -136,7 +126,6 @@ drop materialized view mv1_n2; -- EXAMPLE 5 create materialized view mv1_n2 as select name, salary from emps_n3 group by name, salary; -analyze table mv1_n2 compute statistics for columns; explain select name, salary from emps_n3 group by name, salary; @@ -148,7 +137,6 @@ drop materialized view mv1_n2; -- EXAMPLE 6 create materialized view mv1_n2 as select name, salary from emps_n3 group by name, salary; -analyze table mv1_n2 compute statistics for columns; explain select name from emps_n3 group by name; @@ -160,7 +148,6 @@ drop materialized view mv1_n2; -- EXAMPLE 7 create materialized view mv1_n2 as select name, salary from emps_n3 where deptno = 10 group by name, salary; -analyze table mv1_n2 compute statistics for columns; explain select name from emps_n3 where deptno = 10 group by name; @@ -173,7 +160,6 @@ drop materialized view mv1_n2; create materialized view mv1_n2 as select name, salary, count(*) as c, sum(empid) as s from emps_n3 group by name, salary; -analyze table mv1_n2 compute statistics for columns; explain select name from emps_n3 group by name; diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_10.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_10.q index 9542792316..ff593a797b 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_10.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_10.q @@ -14,13 +14,11 @@ create table emps_n10 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n10 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (110, 10, 'Bill', 10000, 250); -analyze table emps_n10 compute statistics for columns; -- EXAMPLE 1 create materialized view mv1_n10 as select deptno, sum(salary), count(salary) as a from emps_n10 group by deptno; -analyze table mv1_n10 compute statistics for columns; explain select deptno, avg(salary) as a @@ -35,7 +33,6 @@ drop materialized view mv1_n10; create materialized view mv1_n10 as select salary, sum(salary), count(salary) as a from emps_n10 group by salary; -analyze table mv1_n10 compute statistics for columns; explain select salary, avg(salary) as a @@ -50,7 +47,6 @@ drop materialized view mv1_n10; create materialized view mv1_n10 as select salary, sum(salary), count(salary) as a from emps_n10 where salary > 0 group by salary; -analyze table mv1_n10 compute statistics for columns; explain select salary, avg(salary) as a @@ -71,12 +67,10 @@ create table emps_n10_2 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n10_2 values (100, 10, 'Bill', 1, 1000), (200, 20, 'Eric', 2, 500), (150, 10, 'Sebastian', 2, null), (110, 10, 'Theodore', 3, 250), (110, 10, 'Bill', 0, 250); -analyze table emps_n10_2 compute statistics for columns; create materialized view mv1_n10 as select salary, sum(salary), count(salary) as a from emps_n10_2 where salary > 0 group by salary; -analyze table mv1_n10 compute statistics for columns; explain select avg(salary) diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_2.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_2.q index 3a447fc187..feba84ed8b 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_2.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_2.q @@ -15,7 +15,6 @@ create table emps_n0 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n0 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (110, 10, 'Bill', 10000, 250); -analyze table emps_n0 compute statistics for columns; create table depts_n0 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n0 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n0 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n0 compute statistics for columns; create table dependents_n0 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n0 values (10, 'Michael'), (10, 'Jane'); -analyze table dependents_n0 compute statistics for columns; create table locations_n0 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n0 values (10, 'San Francisco'), (10, 'San Diego'); -analyze table locations_n0 compute statistics for columns; alter table emps_n0 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n0 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -52,7 +48,6 @@ create materialized view mv1_n0 as select empid, depts_n0.deptno from emps_n0 join depts_n0 using (deptno) where depts_n0.deptno > 10 group by empid, depts_n0.deptno; -analyze table mv1_n0 compute statistics for columns; explain select empid from emps_n0 @@ -70,7 +65,6 @@ create materialized view mv1_n0 as select depts_n0.deptno, empid from depts_n0 join emps_n0 using (deptno) where depts_n0.deptno > 10 group by empid, depts_n0.deptno; -analyze table mv1_n0 compute statistics for columns; explain select empid from emps_n0 @@ -88,7 +82,6 @@ create materialized view mv1_n0 as select empid, depts_n0.deptno from emps_n0 join depts_n0 using (deptno) where emps_n0.deptno > 10 group by empid, depts_n0.deptno; -analyze table mv1_n0 compute statistics for columns; explain select empid from emps_n0 @@ -106,7 +99,6 @@ create materialized view mv1_n0 as select depts_n0.deptno, emps_n0.empid from depts_n0 join emps_n0 using (deptno) where emps_n0.empid > 10 group by depts_n0.deptno, emps_n0.empid; -analyze table mv1_n0 compute statistics for columns; explain select depts_n0.deptno from depts_n0 @@ -124,7 +116,6 @@ create materialized view mv1_n0 as select depts_n0.deptno, emps_n0.empid from depts_n0 join emps_n0 using (deptno) where emps_n0.empid > 10 group by depts_n0.deptno, emps_n0.empid; -analyze table mv1_n0 compute statistics for columns; explain select depts_n0.deptno from depts_n0 @@ -143,7 +134,6 @@ select depts_n0.name, dependents_n0.name as name2, emps_n0.deptno, depts_n0.dept from depts_n0, dependents_n0, emps_n0 where depts_n0.deptno > 10 group by depts_n0.name, dependents_n0.name, emps_n0.deptno, depts_n0.deptno, dependents_n0.empid; -analyze table mv1_n0 compute statistics for columns; explain select dependents_n0.empid diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_3.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_3.q index 0823f59394..ff02ceb1fa 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_3.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_3.q @@ -15,7 +15,6 @@ create table emps_n9 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n9 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (120, 10, 'Theodore', 10000, 250); -analyze table emps_n9 compute statistics for columns; create table depts_n7 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n7 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n7 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n7 compute statistics for columns; create table dependents_n5 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n5 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n5 compute statistics for columns; create table locations_n5 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n5 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n5 compute statistics for columns; alter table emps_n9 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n7 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -51,7 +47,6 @@ alter table depts_n7 add constraint fk2 foreign key (locationid) references loca create materialized view mv1_n5 as select empid deptno from emps_n9 join depts_n7 using (deptno); -analyze table mv1_n5 compute statistics for columns; explain select empid deptno from emps_n9 @@ -66,7 +61,6 @@ drop materialized view mv1_n5; create materialized view mv1_n5 as select cast(empid as BIGINT) from emps_n9 join depts_n7 using (deptno); -analyze table mv1_n5 compute statistics for columns; explain select empid deptno from emps_n9 @@ -81,7 +75,6 @@ drop materialized view mv1_n5; create materialized view mv1_n5 as select cast(empid as BIGINT) from emps_n9 join depts_n7 using (deptno); -analyze table mv1_n5 compute statistics for columns; explain select empid deptno from emps_n9 @@ -97,7 +90,6 @@ create materialized view mv1_n5 as select depts_n7.name from emps_n9 join depts_n7 on (emps_n9.deptno = depts_n7.deptno); -analyze table mv1_n5 compute statistics for columns; explain select dependents_n5.empid diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_4.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_4.q index 6724cec771..08c1956e21 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_4.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_4.q @@ -15,7 +15,6 @@ create table emps_n5 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n5 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps_n5 compute statistics for columns; create table depts_n4 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n4 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n4 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n4 compute statistics for columns; create table dependents_n3 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n3 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n3 compute statistics for columns; create table locations_n3 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n3 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n3 compute statistics for columns; alter table emps_n5 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n4 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -52,7 +48,6 @@ alter table depts_n4 add constraint fk2 foreign key (locationid) references loca create materialized view mv1_n3 as select name, salary, count(*) as c, sum(empid) as s from emps_n5 group by name, salary; -analyze table mv1_n3 compute statistics for columns; explain select name, count(*) as c, sum(empid) as s @@ -67,7 +62,6 @@ drop materialized view mv1_n3; create materialized view mv1_n3 as select name, salary, count(*) as c, sum(empid) as s from emps_n5 group by name, salary; -analyze table mv1_n3 compute statistics for columns; explain select salary, name, sum(empid) as s, count(*) as c @@ -84,7 +78,6 @@ create materialized view mv1_n3 as select empid, emps_n5.deptno, count(*) as c, sum(empid) as s from emps_n5 join depts_n4 using (deptno) group by empid, emps_n5.deptno; -analyze table mv1_n3 compute statistics for columns; explain select depts_n4.deptno, count(*) as c, sum(empid) as s @@ -102,7 +95,6 @@ create materialized view mv1_n3 as select empid, emps_n5.deptno, count(*) as c, sum(empid) as s from emps_n5 join depts_n4 using (deptno) where emps_n5.deptno >= 10 group by empid, emps_n5.deptno; -analyze table mv1_n3 compute statistics for columns; explain select depts_n4.deptno, sum(empid) as s @@ -120,7 +112,6 @@ create materialized view mv1_n3 as select empid, depts_n4.deptno, count(*) + 1 as c, sum(empid) as s from emps_n5 join depts_n4 using (deptno) where depts_n4.deptno >= 10 group by empid, depts_n4.deptno; -analyze table mv1_n3 compute statistics for columns; explain select depts_n4.deptno, sum(empid) + 1 as s @@ -139,7 +130,6 @@ select depts_n4.name, sum(salary) as s from emps_n5 join depts_n4 on (emps_n5.deptno = depts_n4.deptno) group by depts_n4.name; -analyze table mv1_n3 compute statistics for columns; explain select dependents_n3.empid, sum(salary) as s @@ -162,7 +152,6 @@ select dependents_n3.empid, emps_n5.deptno, count(distinct salary) as s from emps_n5 join dependents_n3 on (emps_n5.empid = dependents_n3.empid) group by dependents_n3.empid, emps_n5.deptno; -analyze table mv1_n3 compute statistics for columns; explain select emps_n5.deptno, count(distinct salary) as s diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_5.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_5.q index d87928c073..1e9a9b99cf 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_5.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_5.q @@ -15,7 +15,6 @@ create table emps_n2 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n2 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps_n2 compute statistics for columns; create table depts_n1 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n1 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n1 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n1 compute statistics for columns; create table dependents_n1 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n1 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n1 compute statistics for columns; create table locations_n1 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n1 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n1 compute statistics for columns; alter table emps_n2 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n1 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -54,7 +50,6 @@ alter table depts_n1 change column locationid locationid int constraint nn2 not -- EXAMPLE 8 create materialized view mv1_n1 as select name, deptno, salary from emps_n2 where deptno > 15 group by name, deptno, salary; -analyze table mv1_n1 compute statistics for columns; explain select name from emps_n2 where deptno >= 20 group by name; @@ -67,7 +62,6 @@ drop materialized view mv1_n1; create materialized view mv1_n1 as select name, deptno, salary, count(*) as c, sum(empid) as s from emps_n2 where deptno >= 15 group by name, deptno, salary; -analyze table mv1_n1 compute statistics for columns; explain select name, sum(empid) as s @@ -87,7 +81,6 @@ join locations_n1 on (locations_n1.name = dependents_n1.name) join emps_n2 on (emps_n2.deptno = depts_n1.deptno) where depts_n1.deptno > 10 and depts_n1.deptno < 20 group by depts_n1.deptno, dependents_n1.empid; -analyze table mv1_n1 compute statistics for columns; explain select dependents_n1.empid @@ -113,7 +106,6 @@ create materialized view mv1_n1 as select empid, depts_n1.deptno, count(*) as c, sum(empid) as s from emps_n2 join depts_n1 using (deptno) group by empid, depts_n1.deptno; -analyze table mv1_n1 compute statistics for columns; explain select deptno from emps_n2 group by deptno; @@ -127,7 +119,6 @@ create materialized view mv1_n1 as select empid, depts_n1.deptno, count(*) as c, sum(empid) as s from emps_n2 join depts_n1 using (deptno) group by empid, depts_n1.deptno; -analyze table mv1_n1 compute statistics for columns; explain select deptno, empid, sum(empid) as s, count(*) as c @@ -144,7 +135,6 @@ select dependents_n1.empid, emps_n2.deptno, sum(salary) as s from emps_n2 join dependents_n1 on (emps_n2.empid = dependents_n1.empid) group by dependents_n1.empid, emps_n2.deptno; -analyze table mv1_n1 compute statistics for columns; explain select dependents_n1.empid, sum(salary) as s @@ -167,7 +157,6 @@ select dependents_n1.empid, emps_n2.deptno, sum(salary) as s from emps_n2 join dependents_n1 on (emps_n2.empid = dependents_n1.empid) group by dependents_n1.empid, emps_n2.deptno; -analyze table mv1_n1 compute statistics for columns; explain select depts_n1.name, sum(salary) as s @@ -190,7 +179,6 @@ select a.empid deptno from (select * from emps_n2 where empid = 1) a join depts_n1 on (a.deptno = depts_n1.deptno) join dependents_n1 on (a.empid = dependents_n1.empid); -analyze table mv1_n1 compute statistics for columns; explain select a.empid from @@ -209,7 +197,6 @@ select a.empid, a.deptno from (select * from emps_n2 where empid = 1) a join depts_n1 on (a.deptno = depts_n1.deptno) join dependents_n1 on (a.empid = dependents_n1.empid); -analyze table mv1_n1 compute statistics for columns; explain select a.empid from @@ -227,7 +214,6 @@ create materialized view mv1_n1 as select empid deptno from (select * from emps_n2 where empid = 1) a join depts_n1 on (a.deptno = depts_n1.deptno); -analyze table mv1_n1 compute statistics for columns; explain select empid from emps_n2 where empid = 1; @@ -242,7 +228,6 @@ select emps_n2.empid, emps_n2.deptno from emps_n2 join depts_n1 on (emps_n2.deptno = depts_n1.deptno) join dependents_n1 on (emps_n2.empid = dependents_n1.empid) where emps_n2.empid = 1; -analyze table mv1_n1 compute statistics for columns; explain select emps_n2.empid from emps_n2 @@ -262,7 +247,6 @@ join depts_n1 a on (emps_n2.deptno=a.deptno) join depts_n1 b on (emps_n2.deptno=b.deptno) join dependents_n1 on (emps_n2.empid = dependents_n1.empid) where emps_n2.empid = 1; -analyze table mv1_n1 compute statistics for columns; explain select emps_n2.empid from emps_n2 @@ -282,7 +266,6 @@ join depts_n1 a on (emps_n2.deptno=a.deptno) join depts_n1 b on (emps_n2.deptno=b.deptno) join dependents_n1 on (emps_n2.empid = dependents_n1.empid) where emps_n2.name = 'Sebastian'; -analyze table mv1_n1 compute statistics for columns; explain select emps_n2.empid from emps_n2 diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_6.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_6.q index 23fc3c14ce..5ff60bcb5e 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_6.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_6.q @@ -15,7 +15,6 @@ create table emps ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps compute statistics for columns; create table depts ( deptno int, @@ -23,21 +22,18 @@ create table depts ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts compute statistics for columns; create table dependents ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents values (10, 'Michael'), (20, 'Jane'); -analyze table dependents compute statistics for columns; create table locations ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations compute statistics for columns; alter table emps add constraint pk1 primary key (empid) disable novalidate rely; alter table depts add constraint pk2 primary key (deptno) disable novalidate rely; @@ -55,7 +51,6 @@ alter table depts change column locationid locationid int constraint nn2 not nul create materialized view mv1 as select name, deptno, salary, count(*) + 1 as c, sum(empid) as s from emps where deptno >= 10 group by name, deptno, salary; -analyze table mv1 compute statistics for columns; explain select salary, sum(empid) + 1 as s @@ -70,7 +65,6 @@ drop materialized view mv1; create materialized view mv1 as select name, deptno, salary, count(*) + 1 as c, sum(empid) as s from emps where deptno >= 15 group by name, deptno, salary; -analyze table mv1 compute statistics for columns; explain select salary + 1, sum(empid) + 1 as s @@ -86,7 +80,6 @@ create materialized view mv1 as select depts.name from emps join depts on (emps.deptno = depts.deptno); -analyze table mv1 compute statistics for columns; explain select dependents.empid @@ -106,7 +99,6 @@ create materialized view mv1 as select depts.name from emps join depts on (emps.deptno = depts.deptno); -analyze table mv1 compute statistics for columns; explain select dependents.empid @@ -127,7 +119,6 @@ drop materialized view mv1; create materialized view mv1 as select emps.empid, emps.deptno, emps.name as name1, emps.salary, emps.commission, dependents.name as name2 from emps join dependents on (emps.empid = dependents.empid); -analyze table mv1 compute statistics for columns; explain select emps.empid, dependents.empid, emps.deptno diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_7.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_7.q index 3d1cedc4f5..9a0e3c8f9a 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_7.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_7.q @@ -15,7 +15,6 @@ create table emps_n8 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n8 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250); -analyze table emps_n8 compute statistics for columns; create table depts_n6 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n6 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n6 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n6 compute statistics for columns; create table dependents_n4 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n4 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n4 compute statistics for columns; create table locations_n4 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n4 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n4 compute statistics for columns; alter table emps_n8 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n6 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -60,7 +56,6 @@ join locations_n4 on (locations_n4.name = dependents_n4.name) join emps_n8 on (emps_n8.deptno = depts_n6.deptno) where depts_n6.deptno > 11 group by depts_n6.deptno, dependents_n4.empid; -analyze table mv1_n4 compute statistics for columns; explain select dependents_n4.empid, depts_n6.deptno @@ -90,7 +85,6 @@ join locations_n4 on (locations_n4.name = dependents_n4.name) join emps_n8 on (emps_n8.deptno = depts_n6.deptno) where depts_n6.deptno > 11 and depts_n6.deptno < 19 group by depts_n6.deptno, dependents_n4.empid; -analyze table mv1_n4 compute statistics for columns; explain select dependents_n4.empid, count(emps_n8.salary) + 1 @@ -119,7 +113,6 @@ from depts_n6 join dependents_n4 on (depts_n6.name = dependents_n4.name) join emps_n8 on (emps_n8.deptno = depts_n6.deptno) where depts_n6.deptno >= 10; -analyze table mv1_n4 compute statistics for columns; explain select dependents_n4.empid diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_8.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_8.q index cfcfddce50..919a356115 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_8.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_8.q @@ -18,7 +18,6 @@ stored AS ORC TBLPROPERTIES("transactional"="true"); insert into table source_table_001 values ('2010-10-10', 1, 1, 'env', 1, 1); -analyze table source_table_001 compute statistics for columns; CREATE MATERIALIZED VIEW source_table_001_mv AS SELECT @@ -27,7 +26,6 @@ SUM(A.UP_VOLUME) AS UP_VOLUME_SUM, A.MY_DATE,A.MY_ID2,A.ENVIRONMENT from source_table_001 AS A group by A.MY_ID,A.MY_ID2,A.ENVIRONMENT,A.MY_DATE; -analyze table source_table_001_mv compute statistics for columns; explain diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_9.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_9.q index 18d5cec8f9..40c4117d63 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_9.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_9.q @@ -18,7 +18,6 @@ stored AS ORC TBLPROPERTIES("transactional"="true"); insert into table source_table_001_n0 values ('2010-10-10 00:00:00', 1, 1, 'env', 1, 1); -analyze table source_table_001_n0 compute statistics for columns; CREATE MATERIALIZED VIEW source_table_001_mv_n0 AS SELECT @@ -27,7 +26,6 @@ SUM(A.UP_VOLUME) AS UP_VOLUME_SUM, A.MY_ID,A.MY_DATE,A.MY_ID2,A.ENVIRONMENT from source_table_001_n0 AS A group by A.MY_ID,A.MY_ID2,A.ENVIRONMENT,A.MY_DATE; -analyze table source_table_001_mv_n0 compute statistics for columns; explain select @@ -45,7 +43,6 @@ SUM(A.UP_VOLUME) AS UP_VOLUME_SUM, A.MY_ID,FLOOR(A.MY_DATE to hour),A.MY_ID2,A.ENVIRONMENT from source_table_001_n0 AS A group by A.MY_ID,A.MY_ID2,A.ENVIRONMENT,FLOOR(A.MY_DATE to hour); -analyze table source_table_001_mv_n0 compute statistics for columns; explain select diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_empty.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_empty.q index 9ae1d4e81b..db56a386c1 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_empty.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_empty.q @@ -13,7 +13,6 @@ create table emps_mv_rewrite_empty ( salary float, commission int) stored as orc TBLPROPERTIES ('transactional'='true'); -analyze table emps_mv_rewrite_empty compute statistics for columns; create materialized view emps_mv_rewrite_empty_mv1 as select * from emps_mv_rewrite_empty where empid < 150; diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt.q index 8de9c7087a..353cef8dbb 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt.q @@ -15,7 +15,6 @@ create table emps_n30 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n30 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps_n30 compute statistics for columns; create table depts_n20 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n20 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n20 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n20 compute statistics for columns; create table dependents_n20 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n20 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n20 compute statistics for columns; create table locations_n20 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n20 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n20 compute statistics for columns; alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n20 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -51,7 +47,6 @@ alter table depts_n20 add constraint fk2 foreign key (locationid) references loc create materialized view mv1_n20 as select deptno, name, salary, commission from emps_n30; -analyze table mv1_n20 compute statistics for columns; explain select emps_n30.name, emps_n30.salary, emps_n30.commission @@ -69,7 +64,6 @@ create materialized view mv1_n20 as select empid, emps_n30.deptno, count(*) as c, sum(empid) as s from emps_n30 join depts_n20 using (deptno) group by empid, emps_n30.deptno; -analyze table mv1_n20 compute statistics for columns; explain select depts_n20.deptno, count(*) as c, sum(empid) as s @@ -88,7 +82,6 @@ select dependents_n20.empid, emps_n30.deptno, sum(salary) as s from emps_n30 join dependents_n20 on (emps_n30.empid = dependents_n20.empid) group by dependents_n20.empid, emps_n30.deptno; -analyze table mv1_n20 compute statistics for columns; explain select dependents_n20.empid, sum(salary) as s @@ -109,7 +102,6 @@ drop materialized view mv1_n20; create materialized view mv1_n20 as select emps_n30.empid, emps_n30.deptno, emps_n30.name as name1, emps_n30.salary, emps_n30.commission, dependents_n20.name as name2 from emps_n30 join dependents_n20 on (emps_n30.empid = dependents_n20.empid); -analyze table mv1_n20 compute statistics for columns; explain select emps_n30.empid, dependents_n20.empid, emps_n30.deptno diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt_2.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt_2.q index a1372301fe..a5bb2c0512 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt_2.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_no_join_opt_2.q @@ -15,7 +15,6 @@ create table emps_n30 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n30 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps_n30 compute statistics for columns; create table depts_n20 ( deptno int, @@ -23,21 +22,18 @@ create table depts_n20 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n20 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n20 compute statistics for columns; create table dependents_n20 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n20 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n20 compute statistics for columns; create table locations_n20 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n20 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n20 compute statistics for columns; alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n20 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -50,7 +46,6 @@ alter table depts_n20 add constraint fk2 foreign key (locationid) references loc -- EXAMPLE 1 create materialized view mv1_part_n2 partitioned on (deptno) as select * from emps_n30 where empid < 150; -analyze table mv1_part_n2 compute statistics for columns; explain select * @@ -67,7 +62,6 @@ drop materialized view mv1_part_n2; create materialized view mv1_part_n2 partitioned on (deptno) as select deptno, name, salary, commission from emps_n30; -analyze table mv1_part_n2 compute statistics for columns; explain select emps_n30.name, emps_n30.salary, emps_n30.commission @@ -83,7 +77,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 4 create materialized view mv1_part_n2 partitioned on (deptno) as select * from emps_n30 where empid < 200; -analyze table mv1_part_n2 compute statistics for columns; explain select * from emps_n30 where empid > 120 @@ -97,7 +90,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 5 create materialized view mv1_part_n2 partitioned on (name) as select name, salary from emps_n30 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name, salary from emps_n30 group by name, salary; @@ -109,7 +101,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 6 create materialized view mv1_part_n2 partitioned on (name) as select name, salary from emps_n30 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name from emps_n30 group by name; @@ -121,7 +112,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 7 create materialized view mv1_part_n2 partitioned on (name) as select name, salary from emps_n30 where deptno = 10 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name from emps_n30 where deptno = 10 group by name; @@ -134,7 +124,6 @@ drop materialized view mv1_part_n2; create materialized view mv1_part_n2 partitioned on (name) as select name, salary, count(*) as c, sum(empid) as s from emps_n30 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name from emps_n30 group by name; diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_1.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_1.q index e6980c07f1..3fecac15bb 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_1.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_1.q @@ -16,7 +16,6 @@ create table emps_n30 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n30 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (120, 10, 'Bill', 10000, 250); -analyze table emps_n30 compute statistics for columns; create table depts_n20 ( deptno int, @@ -24,21 +23,18 @@ create table depts_n20 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n20 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n20 compute statistics for columns; create table dependents_n20 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n20 values (10, 'Michael'), (20, 'Jane'); -analyze table dependents_n20 compute statistics for columns; create table locations_n20 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n20 values (10, 'San Francisco'), (20, 'San Diego'); -analyze table locations_n20 compute statistics for columns; alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n20 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -51,7 +47,6 @@ alter table depts_n20 add constraint fk2 foreign key (locationid) references loc -- EXAMPLE 1 create materialized view mv1_part_n2 partitioned on (deptno) as select * from emps_n30 where empid < 150; -analyze table mv1_part_n2 compute statistics for columns; explain select * @@ -68,7 +63,6 @@ drop materialized view mv1_part_n2; create materialized view mv1_part_n2 partitioned on (deptno) as select deptno, name, salary, commission from emps_n30; -analyze table mv1_part_n2 compute statistics for columns; explain select emps_n30.name, emps_n30.salary, emps_n30.commission @@ -84,7 +78,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 4 create materialized view mv1_part_n2 partitioned on (deptno) as select * from emps_n30 where empid < 200; -analyze table mv1_part_n2 compute statistics for columns; explain select * from emps_n30 where empid > 120 @@ -98,7 +91,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 5 create materialized view mv1_part_n2 partitioned on (name) as select name, salary from emps_n30 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name, salary from emps_n30 group by name, salary; @@ -110,7 +102,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 6 create materialized view mv1_part_n2 partitioned on (name) as select name, salary from emps_n30 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name from emps_n30 group by name; @@ -122,7 +113,6 @@ drop materialized view mv1_part_n2; -- EXAMPLE 7 create materialized view mv1_part_n2 partitioned on (name) as select name, salary from emps_n30 where deptno = 10 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name from emps_n30 where deptno = 10 group by name; @@ -135,7 +125,6 @@ drop materialized view mv1_part_n2; create materialized view mv1_part_n2 partitioned on (name) as select name, salary, count(*) as c, sum(empid) as s from emps_n30 group by name, salary; -analyze table mv1_part_n2 compute statistics for columns; explain select name from emps_n30 group by name; diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_2.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_2.q index b2e6ebd695..8d6df04b7c 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_2.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_part_2.q @@ -16,7 +16,6 @@ create table emps_n00 ( stored as orc TBLPROPERTIES ('transactional'='true'); insert into emps_n00 values (100, 10, 'Bill', 10000, 1000), (200, 20, 'Eric', 8000, 500), (150, 10, 'Sebastian', 7000, null), (110, 10, 'Theodore', 10000, 250), (110, 10, 'Bill', 10000, 250); -analyze table emps_n00 compute statistics for columns; create table depts_n00 ( deptno int, @@ -24,21 +23,18 @@ create table depts_n00 ( locationid int) stored as orc TBLPROPERTIES ('transactional'='true'); insert into depts_n00 values (10, 'Sales', 10), (30, 'Marketing', null), (20, 'HR', 20); -analyze table depts_n00 compute statistics for columns; create table dependents_n00 ( empid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into dependents_n00 values (10, 'Michael'), (10, 'Jane'); -analyze table dependents_n00 compute statistics for columns; create table locations_n00 ( locationid int, name varchar(256)) stored as orc TBLPROPERTIES ('transactional'='true'); insert into locations_n00 values (10, 'San Francisco'), (10, 'San Diego'); -analyze table locations_n00 compute statistics for columns; alter table emps_n00 add constraint pk1 primary key (empid) disable novalidate rely; alter table depts_n00 add constraint pk2 primary key (deptno) disable novalidate rely; @@ -53,7 +49,6 @@ create materialized view mv1_part_n0 partitioned on (deptno) as select empid, depts_n00.deptno as deptno from emps_n00 join depts_n00 using (deptno) where depts_n00.deptno > 10 group by empid, depts_n00.deptno; -analyze table mv1_part_n0 compute statistics for columns; explain select empid from emps_n00 @@ -71,7 +66,6 @@ create materialized view mv1_part_n0 partitioned on (deptno) as select depts_n00.deptno as deptno, empid from depts_n00 join emps_n00 using (deptno) where depts_n00.deptno > 10 group by empid, depts_n00.deptno; -analyze table mv1_part_n0 compute statistics for columns; explain select empid from emps_n00 @@ -89,7 +83,6 @@ create materialized view mv1_part_n0 partitioned on (deptno) as select empid, depts_n00.deptno as deptno from emps_n00 join depts_n00 using (deptno) where emps_n00.deptno > 10 group by empid, depts_n00.deptno; -analyze table mv1_part_n0 compute statistics for columns; explain select empid from emps_n00 @@ -107,7 +100,6 @@ create materialized view mv1_part_n0 partitioned on (deptno) as select depts_n00.deptno as deptno, emps_n00.empid from depts_n00 join emps_n00 using (deptno) where emps_n00.empid > 10 group by depts_n00.deptno, emps_n00.empid; -analyze table mv1_part_n0 compute statistics for columns; explain select depts_n00.deptno from depts_n00 @@ -125,7 +117,6 @@ create materialized view mv1_part_n0 partitioned on (deptno) as select depts_n00.deptno as deptno, emps_n00.empid from depts_n00 join emps_n00 using (deptno) where emps_n00.empid > 10 group by depts_n00.deptno, emps_n00.empid; -analyze table mv1_part_n0 compute statistics for columns; explain select depts_n00.deptno from depts_n00 @@ -144,7 +135,6 @@ select depts_n00.name, dependents_n00.name as name2, emps_n00.deptno, depts_n00. from depts_n00, dependents_n00, emps_n00 where depts_n00.deptno > 10 group by depts_n00.name, dependents_n00.name, emps_n00.deptno, depts_n00.deptno, dependents_n00.empid; -analyze table mv1_part_n0 compute statistics for columns; explain select dependents_n00.empid diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb.q index aed5bdbffd..2bd92bd25b 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb.q @@ -202,12 +202,6 @@ TBLPROPERTIES ('transactional'='true'); INSERT INTO `lineorder_n0` SELECT * FROM `lineorder_ext_n0`; -analyze table customer_n1 compute statistics for columns; -analyze table dates_n0 compute statistics for columns; -analyze table ssb_part_n0 compute statistics for columns; -analyze table supplier_n0 compute statistics for columns; -analyze table lineorder_n0 compute statistics for columns; - CREATE MATERIALIZED VIEW `ssb_mv_n0` AS SELECT diff --git a/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb_2.q b/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb_2.q index 0982b66ad7..1f4621ede5 100644 --- a/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb_2.q +++ b/ql/src/test/queries/clientpositive/materialized_view_rewrite_ssb_2.q @@ -202,12 +202,6 @@ TBLPROPERTIES ('transactional'='true'); INSERT INTO `lineorder` SELECT * FROM `lineorder_ext`; -analyze table customer_n0 compute statistics for columns; -analyze table dates compute statistics for columns; -analyze table ssb_part compute statistics for columns; -analyze table supplier compute statistics for columns; -analyze table lineorder compute statistics for columns; - CREATE MATERIALIZED VIEW `ssb_mv` AS SELECT diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create.q.out index 14bf676db9..15404830d6 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create.q.out @@ -45,7 +45,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"b\":\"true\",\"c\":\"true\"}} bucketing_version 2 numFiles 1 numRows 5 @@ -107,7 +107,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 numFiles 1 numRows 5 @@ -243,7 +243,7 @@ PREHOOK: query: show tblproperties cmv_mat_view5 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view5 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true","c":"true","d":"true","f":"true"}} bucketing_version 2 key value numFiles 1 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite.q.out index 69bdbeace2..6bf9ee5e54 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n10.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n10.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n10.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n10.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n10 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n10 -PREHOOK: Output: default@cmv_basetable_n10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n10 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n10 -POSTHOOK: Output: default@cmv_basetable_n10 -#### A masked pattern was here #### PREHOOK: query: create materialized view cmv_mat_view_n10 as select a, b, c from cmv_basetable_n10 where a = 2 PREHOOK: type: CREATE_MATERIALIZED_VIEW @@ -64,7 +54,7 @@ PREHOOK: query: show tblproperties cmv_mat_view_n10 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view_n10 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true","c":"true"}} bucketing_version 2 numFiles 1 numFilesErasureCoded 0 @@ -98,7 +88,7 @@ PREHOOK: query: show tblproperties cmv_mat_view2_n4 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view2_n4 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","c":"true"}} bucketing_version 2 numFiles 1 numFilesErasureCoded 0 @@ -354,10 +344,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: default.cmv_mat_view2_n4 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE value expressions: a (type: int), c (type: decimal(10,2)) Execution mode: llap LLAP IO: all inputs @@ -371,14 +361,14 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 2 Data size: 458 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 456 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col0 (type: int), _col1 (type: decimal(10,2)), _col0 (type: int), _col5 (type: decimal(10,2)) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 458 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 464 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 458 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 464 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_2.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_2.q.out index ce57ec1b7d..82415dee2d 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_2.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_2.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n9.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n9.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n9.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n9.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n9 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n9 -PREHOOK: Output: default@cmv_basetable_n9 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n9 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n9 -POSTHOOK: Output: default@cmv_basetable_n9 -#### A masked pattern was here #### PREHOOK: query: create materialized view cmv_mat_view_n9 as select b from cmv_basetable_n9 where c > 10.0 group by a, b, c PREHOOK: type: CREATE_MATERIALIZED_VIEW @@ -269,24 +259,24 @@ STAGE PLANS: TableScan alias: default.cmv_mat_view_3 filterExpr: (c > 20) (type: boolean) - Statistics: Num rows: 3 Data size: 1368 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 618 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: (c > 20) (type: boolean) - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), b (type: varchar(256)) outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: a (type: int), b (type: varchar(256)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 94 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: varchar(256)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: varchar(256)) - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 94 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: llap LLAP IO: all inputs Reducer 2 @@ -296,14 +286,14 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: varchar(256)) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 94 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: varchar(256)) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 90 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 90 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -367,21 +357,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: default.cmv_mat_view_4 - Statistics: Num rows: 5 Data size: 1700 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 5 Data size: 450 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: b (type: varchar(256)) outputColumnNames: b - Statistics: Num rows: 5 Data size: 1700 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 5 Data size: 450 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: b (type: varchar(256)) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 1700 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: varchar(256)) sort order: + Map-reduce partition columns: _col0 (type: varchar(256)) - Statistics: Num rows: 5 Data size: 1700 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: llap LLAP IO: all inputs Reducer 2 @@ -391,10 +381,10 @@ STAGE PLANS: keys: KEY._col0 (type: varchar(256)) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -445,16 +435,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n4.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n4.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n4.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n4.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n4 -PREHOOK: Output: default@cmv_basetable_2_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n4 -POSTHOOK: Output: default@cmv_basetable_2_n4 -#### A masked pattern was here #### PREHOOK: query: create materialized view cmv_mat_view_5 as select cmv_basetable_n9.a, cmv_basetable_2_n4.c from cmv_basetable_n9 join cmv_basetable_2_n4 on (cmv_basetable_n9.a = cmv_basetable_2_n4.a) @@ -570,24 +550,24 @@ STAGE PLANS: TableScan alias: default.cmv_mat_view_5 filterExpr: (c > 10.1) (type: boolean) - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: (c > 10.1) (type: boolean) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int) outputColumnNames: a - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: a (type: int) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: llap LLAP IO: all inputs Reducer 2 @@ -597,10 +577,10 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_3.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_3.q.out index 8fcab1f621..e8d0657129 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_3.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_3.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable -PREHOOK: Output: default@cmv_basetable -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable -POSTHOOK: Output: default@cmv_basetable -#### A masked pattern was here #### PREHOOK: query: create table cmv_basetable_2 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default @@ -62,16 +52,6 @@ POSTHOOK: Lineage: cmv_basetable_2.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2 -PREHOOK: Output: default@cmv_basetable_2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2 -POSTHOOK: Output: default@cmv_basetable_2 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view AS SELECT cmv_basetable.a, cmv_basetable_2.c @@ -97,7 +77,7 @@ POSTHOOK: Output: default@cmv_mat_view STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 + Stage-4 depends on stages: Stage-0, Stage-2 Stage-3 depends on stages: Stage-4 Stage-5 depends on stages: Stage-3 Stage-0 depends on stages: Stage-1 @@ -107,8 +87,9 @@ STAGE PLANS: Tez #### A masked pattern was here #### Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### Vertices: Map 1 @@ -131,7 +112,7 @@ STAGE PLANS: Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: llap LLAP IO: may be used (ACID table) - Map 4 + Map 5 Map Operator Tree: TableScan alias: cmv_basetable_2 @@ -189,6 +170,34 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.cmv_mat_view + Select Operator + expressions: _col0 (type: int), _col1 (type: decimal(10,2)) + outputColumnNames: col1, col2 + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll') + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + sort order: + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-2 Dependency Collection @@ -211,6 +220,10 @@ STAGE PLANS: Stage: Stage-3 Stats Work Basic Stats Work: + Column Stats Desc: + Columns: a, c + Column Types: int, decimal(10,2) + Table: default.cmv_mat_view Stage: Stage-5 Materialized View Work @@ -313,16 +326,6 @@ POSTHOOK: Lineage: cmv_basetable_2.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2 -PREHOOK: Output: default@cmv_basetable_2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2 -POSTHOOK: Output: default@cmv_basetable_2 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN SELECT cmv_basetable.a FROM cmv_basetable join cmv_basetable_2 ON (cmv_basetable.a = cmv_basetable_2.a) @@ -518,10 +521,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2 - filterExpr: ((c > 10) and (ROW__ID.writeid > 2) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 1) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 2) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 1) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) @@ -539,21 +542,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: default.cmv_mat_view - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: _col0 (type: int), _col1 (type: decimal(10,2)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: decimal(10,2)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: llap LLAP IO: all inputs Reducer 2 @@ -589,12 +592,12 @@ STAGE PLANS: keys: _col0 (type: int), _col1 (type: decimal(10,2)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: decimal(10,2)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reducer 5 Execution mode: llap Reduce Operator Tree: @@ -602,10 +605,10 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: decimal(10,2)) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat @@ -614,15 +617,15 @@ STAGE PLANS: Select Operator expressions: _col0 (type: int), _col1 (type: decimal(10,2)) outputColumnNames: a, c - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: compute_stats(a, 'hll'), compute_stats(c, 'hll') mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: struct), _col1 (type: struct) Reducer 6 Execution mode: llap @@ -631,10 +634,10 @@ STAGE PLANS: aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -976,16 +979,6 @@ POSTHOOK: Lineage: cmv_irrelevant_table.a SCRIPT [] POSTHOOK: Lineage: cmv_irrelevant_table.b SCRIPT [] POSTHOOK: Lineage: cmv_irrelevant_table.c SCRIPT [] POSTHOOK: Lineage: cmv_irrelevant_table.d SCRIPT [] -PREHOOK: query: analyze table cmv_irrelevant_table compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_irrelevant_table -PREHOOK: Output: default@cmv_irrelevant_table -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_irrelevant_table compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_irrelevant_table -POSTHOOK: Output: default@cmv_irrelevant_table -#### A masked pattern was here #### PREHOOK: query: EXPLAIN SELECT cmv_basetable.a FROM cmv_basetable join cmv_basetable_2 ON (cmv_basetable.a = cmv_basetable_2.a) diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_4.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_4.q.out index 8fe194e97b..e107898032 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_4.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_4.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n5.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n5.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n5.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n5.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n5 -PREHOOK: Output: default@cmv_basetable_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n5 -POSTHOOK: Output: default@cmv_basetable_n5 -#### A masked pattern was here #### PREHOOK: query: create table cmv_basetable_2_n2 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default @@ -62,16 +52,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n2.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n2.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n2.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n2.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n2 -PREHOOK: Output: default@cmv_basetable_2_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n2 -POSTHOOK: Output: default@cmv_basetable_2_n2 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view_n5 DISABLE REWRITE TBLPROPERTIES ('transactional'='true') AS SELECT cmv_basetable_n5.a, cmv_basetable_2_n2.c, sum(cmv_basetable_2_n2.d) @@ -97,7 +77,7 @@ POSTHOOK: Output: default@cmv_mat_view_n5 STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 + Stage-4 depends on stages: Stage-0, Stage-2 Stage-3 depends on stages: Stage-4 Stage-5 depends on stages: Stage-3 Stage-0 depends on stages: Stage-1 @@ -192,6 +172,27 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.cmv_mat_view_n5 + Write Type: INSERT + Select Operator + expressions: _col0 (type: int), _col1 (type: decimal(10,2)), _col2 (type: bigint) + outputColumnNames: col1, col2, col3 + Statistics: Num rows: 1 Data size: 124 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll') + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1480 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1480 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1480 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-2 Dependency Collection @@ -215,6 +216,10 @@ STAGE PLANS: Stage: Stage-3 Stats Work Basic Stats Work: + Column Stats Desc: + Columns: a, c, _c2 + Column Types: int, decimal(10,2), bigint + Table: default.cmv_mat_view_n5 Stage: Stage-5 Materialized View Work @@ -224,6 +229,7 @@ STAGE PLANS: files: hdfs directory: true #### A masked pattern was here #### + Write Type: INSERT PREHOOK: query: CREATE MATERIALIZED VIEW cmv_mat_view_n5 DISABLE REWRITE TBLPROPERTIES ('transactional'='true') AS SELECT cmv_basetable_n5.a, cmv_basetable_2_n2.c, sum(cmv_basetable_2_n2.d) @@ -245,16 +251,6 @@ POSTHOOK: Input: default@cmv_basetable_2_n2 POSTHOOK: Input: default@cmv_basetable_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@cmv_mat_view_n5 -PREHOOK: query: analyze table cmv_mat_view_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_mat_view_n5 -PREHOOK: Output: default@cmv_mat_view_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_mat_view_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_mat_view_n5 -POSTHOOK: Output: default@cmv_mat_view_n5 -#### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED cmv_mat_view_n5 PREHOOK: type: DESCTABLE PREHOOK: Input: default@cmv_mat_view_n5 @@ -275,10 +271,10 @@ Table Type: MATERIALIZED_VIEW Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"_c2\":\"true\",\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 - numFiles 2 + numFiles 1 numRows 2 rawDataSize 0 - totalSize 736 + totalSize 809 transactional true transactional_properties default #### A masked pattern was here #### @@ -457,16 +453,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n2.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n2.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n2.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n2.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n2 -PREHOOK: Output: default@cmv_basetable_2_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n2 -POSTHOOK: Output: default@cmv_basetable_2_n2 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN ALTER MATERIALIZED VIEW cmv_mat_view_n5 ENABLE REWRITE PREHOOK: type: ALTER_MATERIALIZED_VIEW_REWRITE @@ -515,10 +501,10 @@ Table Type: MATERIALIZED_VIEW Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"_c2\":\"true\",\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 - numFiles 2 + numFiles 1 numRows 2 rawDataSize 0 - totalSize 736 + totalSize 809 transactional true transactional_properties default #### A masked pattern was here #### @@ -766,10 +752,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2_n2 - filterExpr: ((c > 10) and (ROW__ID.writeid > 2) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 1) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 360 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 2) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 1) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)), d (type: int) @@ -983,10 +969,10 @@ Table Type: MATERIALIZED_VIEW Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} bucketing_version 2 - numFiles 3 + numFiles 2 numRows 3 rawDataSize 0 - totalSize 1500 + totalSize 1573 transactional true transactional_properties default #### A masked pattern was here #### @@ -1637,10 +1623,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2_n2 - filterExpr: ((c > 10) and (ROW__ID.writeid > 6) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 4) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 360 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 6) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 4) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 120 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)), d (type: int) diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_5.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_5.q.out index 117593e244..1b0d9e09af 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_5.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_5.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n6.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n6.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n6.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n6.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n6 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n6 -PREHOOK: Output: default@cmv_basetable_n6 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n6 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n6 -POSTHOOK: Output: default@cmv_basetable_n6 -#### A masked pattern was here #### PREHOOK: query: create table cmv_basetable_2_n3 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default @@ -62,16 +52,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n3.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n3.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n3.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n3.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n3 -PREHOOK: Output: default@cmv_basetable_2_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n3 -POSTHOOK: Output: default@cmv_basetable_2_n3 -#### A masked pattern was here #### PREHOOK: query: CREATE MATERIALIZED VIEW cmv_mat_view_n6 TBLPROPERTIES ('transactional'='true') AS SELECT cmv_basetable_n6.a, cmv_basetable_2_n3.c @@ -92,16 +72,6 @@ POSTHOOK: Input: default@cmv_basetable_2_n3 POSTHOOK: Input: default@cmv_basetable_n6 POSTHOOK: Output: database:default POSTHOOK: Output: default@cmv_mat_view_n6 -PREHOOK: query: analyze table cmv_mat_view_n6 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_mat_view_n6 -PREHOOK: Output: default@cmv_mat_view_n6 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_mat_view_n6 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_mat_view_n6 -POSTHOOK: Output: default@cmv_mat_view_n6 -#### A masked pattern was here #### PREHOOK: query: insert into cmv_basetable_2_n3 values (3, 'charlie', 15.8, 1) PREHOOK: type: QUERY @@ -116,16 +86,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n3.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n3.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n3.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n3.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n3 -PREHOOK: Output: default@cmv_basetable_2_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n3 -POSTHOOK: Output: default@cmv_basetable_2_n3 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN SELECT cmv_basetable_n6.a FROM cmv_basetable_n6 join cmv_basetable_2_n3 ON (cmv_basetable_n6.a = cmv_basetable_2_n3.a) @@ -290,10 +250,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2_n3 - filterExpr: ((c > 10) and (ROW__ID.writeid > 2) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 1) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 2) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 1) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) @@ -419,7 +379,7 @@ Table Parameters: numFiles 2 numRows 5 rawDataSize 0 - totalSize 1071 + totalSize 1498 transactional true transactional_properties default #### A masked pattern was here #### @@ -1013,10 +973,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2_n3 - filterExpr: ((c > 10) and (ROW__ID.writeid > 6) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 4) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 6) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 4) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_dummy.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_dummy.q.out index 485a32e86f..734ae685ef 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_dummy.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_dummy.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n0.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n0.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n0.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n0.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n0 -PREHOOK: Output: default@cmv_basetable_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n0 -POSTHOOK: Output: default@cmv_basetable_n0 -#### A masked pattern was here #### PREHOOK: query: create materialized view cmv_mat_view_n0 as select a, b, c from cmv_basetable_n0 where a = 2 PREHOOK: type: CREATE_MATERIALIZED_VIEW @@ -64,7 +54,7 @@ PREHOOK: query: show tblproperties cmv_mat_view_n0 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view_n0 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true","c":"true"}} bucketing_version 2 numFiles 1 numFilesErasureCoded 0 @@ -98,7 +88,7 @@ PREHOOK: query: show tblproperties cmv_mat_view2 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view2 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","c":"true"}} bucketing_version 2 numFiles 1 numFilesErasureCoded 0 @@ -354,10 +344,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: default.cmv_mat_view2 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE value expressions: a (type: int), c (type: decimal(10,2)) Execution mode: llap LLAP IO: all inputs @@ -371,14 +361,14 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 2 Data size: 458 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 456 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col0 (type: int), _col1 (type: decimal(10,2)), _col0 (type: int), _col5 (type: decimal(10,2)) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 458 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 464 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 458 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 464 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_multi_db.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_multi_db.q.out index ecaefad5b6..4bf5c2f972 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_multi_db.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_multi_db.q.out @@ -40,16 +40,6 @@ POSTHOOK: Lineage: cmv_basetable_n7.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n7.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n7.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n7.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n7 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: db1@cmv_basetable_n7 -PREHOOK: Output: db1@cmv_basetable_n7 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n7 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: db1@cmv_basetable_n7 -POSTHOOK: Output: db1@cmv_basetable_n7 -#### A masked pattern was here #### PREHOOK: query: create database db2 PREHOOK: type: CREATEDATABASE PREHOOK: Output: database:db2 @@ -88,7 +78,7 @@ PREHOOK: query: show tblproperties cmv_mat_view_n7 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view_n7 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true","c":"true"}} bucketing_version 2 numFiles 1 numFilesErasureCoded 0 @@ -122,7 +112,7 @@ PREHOOK: query: show tblproperties cmv_mat_view2_n2 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view2_n2 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","c":"true"}} bucketing_version 2 numFiles 1 numFilesErasureCoded 0 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_rebuild_dummy.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_rebuild_dummy.q.out index 32befa5119..261cc5fdf9 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_rebuild_dummy.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_rebuild_dummy.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n1.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n1.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n1.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n1.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n1 -PREHOOK: Output: default@cmv_basetable_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n1 -POSTHOOK: Output: default@cmv_basetable_n1 -#### A masked pattern was here #### PREHOOK: query: create table cmv_basetable_2_n0 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default @@ -62,16 +52,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n0.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n0.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n0.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n0.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n0 -PREHOOK: Output: default@cmv_basetable_2_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n0 -POSTHOOK: Output: default@cmv_basetable_2_n0 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view_n1 AS SELECT cmv_basetable_n1.a, cmv_basetable_2_n0.c @@ -97,7 +77,7 @@ POSTHOOK: Output: default@cmv_mat_view_n1 STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 + Stage-4 depends on stages: Stage-0, Stage-2 Stage-3 depends on stages: Stage-4 Stage-5 depends on stages: Stage-3 Stage-0 depends on stages: Stage-1 @@ -107,8 +87,9 @@ STAGE PLANS: Tez #### A masked pattern was here #### Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### Vertices: Map 1 @@ -131,7 +112,7 @@ STAGE PLANS: Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: vectorized, llap LLAP IO: may be used (ACID table) - Map 4 + Map 5 Map Operator Tree: TableScan alias: cmv_basetable_2_n0 @@ -174,7 +155,7 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reducer 3 - Execution mode: vectorized, llap + Execution mode: llap Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int), KEY._col1 (type: decimal(10,2)) @@ -189,6 +170,34 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.cmv_mat_view_n1 + Select Operator + expressions: _col0 (type: int), _col1 (type: decimal(10,2)) + outputColumnNames: col1, col2 + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll') + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + sort order: + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-2 Dependency Collection @@ -211,6 +220,10 @@ STAGE PLANS: Stage: Stage-3 Stats Work Basic Stats Work: + Column Stats Desc: + Columns: a, c + Column Types: int, decimal(10,2) + Table: default.cmv_mat_view_n1 Stage: Stage-5 Materialized View Work @@ -313,16 +326,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n0.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n0.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n0.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n0.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n0 -PREHOOK: Output: default@cmv_basetable_2_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n0 -POSTHOOK: Output: default@cmv_basetable_2_n0 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN SELECT cmv_basetable_n1.a FROM cmv_basetable_n1 join cmv_basetable_2_n0 ON (cmv_basetable_n1.a = cmv_basetable_2_n0.a) @@ -518,10 +521,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2_n0 - filterExpr: ((c > 10) and (ROW__ID.writeid > 2) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 1) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 2) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 1) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) @@ -539,21 +542,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: default.cmv_mat_view_n1 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: _col0 (type: int), _col1 (type: decimal(10,2)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: decimal(10,2)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: vectorized, llap LLAP IO: all inputs Reducer 2 @@ -589,12 +592,12 @@ STAGE PLANS: keys: _col0 (type: int), _col1 (type: decimal(10,2)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: decimal(10,2)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reducer 5 Execution mode: llap Reduce Operator Tree: @@ -602,10 +605,10 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: decimal(10,2)) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat @@ -614,15 +617,15 @@ STAGE PLANS: Select Operator expressions: _col0 (type: int), _col1 (type: decimal(10,2)) outputColumnNames: a, c - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: compute_stats(a, 'hll'), compute_stats(c, 'hll') mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: struct), _col1 (type: struct) Reducer 6 Execution mode: llap @@ -631,10 +634,10 @@ STAGE PLANS: aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -976,16 +979,6 @@ POSTHOOK: Lineage: cmv_irrelevant_table_n0.a SCRIPT [] POSTHOOK: Lineage: cmv_irrelevant_table_n0.b SCRIPT [] POSTHOOK: Lineage: cmv_irrelevant_table_n0.c SCRIPT [] POSTHOOK: Lineage: cmv_irrelevant_table_n0.d SCRIPT [] -PREHOOK: query: analyze table cmv_irrelevant_table_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_irrelevant_table_n0 -PREHOOK: Output: default@cmv_irrelevant_table_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_irrelevant_table_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_irrelevant_table_n0 -POSTHOOK: Output: default@cmv_irrelevant_table_n0 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN SELECT cmv_basetable_n1.a FROM cmv_basetable_n1 join cmv_basetable_2_n0 ON (cmv_basetable_n1.a = cmv_basetable_2_n0.a) diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window.q.out index 06835a18f0..51be03e487 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n3.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n3.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n3.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n3.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n3 -PREHOOK: Output: default@cmv_basetable_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n3 -POSTHOOK: Output: default@cmv_basetable_n3 -#### A masked pattern was here #### PREHOOK: query: create table cmv_basetable_2_n1 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default @@ -62,16 +52,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n1.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n1.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n1.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n1.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n1 -PREHOOK: Output: default@cmv_basetable_2_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n1 -POSTHOOK: Output: default@cmv_basetable_2_n1 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN CREATE MATERIALIZED VIEW cmv_mat_view_n3 DISABLE REWRITE TBLPROPERTIES('rewriting.time.window'='5min') AS SELECT cmv_basetable_n3.a, cmv_basetable_2_n1.c @@ -97,7 +77,7 @@ POSTHOOK: Output: default@cmv_mat_view_n3 STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 + Stage-4 depends on stages: Stage-0, Stage-2 Stage-3 depends on stages: Stage-4 Stage-5 depends on stages: Stage-3 Stage-0 depends on stages: Stage-1 @@ -107,8 +87,9 @@ STAGE PLANS: Tez #### A masked pattern was here #### Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### Vertices: Map 1 @@ -131,7 +112,7 @@ STAGE PLANS: Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: vectorized, llap LLAP IO: may be used (ACID table) - Map 4 + Map 5 Map Operator Tree: TableScan alias: cmv_basetable_2_n1 @@ -174,7 +155,7 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reducer 3 - Execution mode: vectorized, llap + Execution mode: llap Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int), KEY._col1 (type: decimal(10,2)) @@ -189,6 +170,34 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.cmv_mat_view_n3 + Select Operator + expressions: _col0 (type: int), _col1 (type: decimal(10,2)) + outputColumnNames: col1, col2 + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll') + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + sort order: + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-2 Dependency Collection @@ -212,6 +221,10 @@ STAGE PLANS: Stage: Stage-3 Stats Work Basic Stats Work: + Column Stats Desc: + Columns: a, c + Column Types: int, decimal(10,2) + Table: default.cmv_mat_view_n3 Stage: Stage-5 Materialized View Work @@ -259,7 +272,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 numFiles 2 numRows 2 @@ -439,16 +452,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n1.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n1.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n1.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n1.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n1 -PREHOOK: Output: default@cmv_basetable_2_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n1 -POSTHOOK: Output: default@cmv_basetable_2_n1 -#### A masked pattern was here #### PREHOOK: query: EXPLAIN ALTER MATERIALIZED VIEW cmv_mat_view_n3 ENABLE REWRITE PREHOOK: type: ALTER_MATERIALIZED_VIEW_REWRITE @@ -494,7 +497,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 numFiles 2 numRows 2 @@ -638,10 +641,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: cmv_basetable_2_n1 - filterExpr: ((c > 10) and (ROW__ID.writeid > 2) and a is not null) (type: boolean) + filterExpr: ((c > 10) and (ROW__ID.writeid > 1) and a is not null) (type: boolean) Statistics: Num rows: 3 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator - predicate: ((ROW__ID.writeid > 2) and (c > 10) and a is not null) (type: boolean) + predicate: ((ROW__ID.writeid > 1) and (c > 10) and a is not null) (type: boolean) Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) @@ -659,21 +662,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: default.cmv_mat_view_n3 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: a (type: int), c (type: decimal(10,2)) outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 232 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: _col0 (type: int), _col1 (type: decimal(10,2)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: decimal(10,2)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: vectorized, llap LLAP IO: all inputs Reducer 2 @@ -709,12 +712,12 @@ STAGE PLANS: keys: _col0 (type: int), _col1 (type: decimal(10,2)) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: decimal(10,2)) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: decimal(10,2)) - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Reducer 5 Execution mode: llap Reduce Operator Tree: @@ -722,10 +725,10 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: decimal(10,2)) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat @@ -734,15 +737,15 @@ STAGE PLANS: Select Operator expressions: _col0 (type: int), _col1 (type: decimal(10,2)) outputColumnNames: a, c - Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: compute_stats(a, 'hll'), compute_stats(c, 'hll') mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: struct), _col1 (type: struct) Reducer 6 Execution mode: llap @@ -751,10 +754,10 @@ STAGE PLANS: aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 1088 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window_2.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window_2.q.out index 24bd61d7f8..b382fa49b5 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window_2.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_create_rewrite_time_window_2.q.out @@ -28,16 +28,6 @@ POSTHOOK: Lineage: cmv_basetable_n100.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n100.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n100.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_n100.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_n100 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_n100 -PREHOOK: Output: default@cmv_basetable_n100 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_n100 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_n100 -POSTHOOK: Output: default@cmv_basetable_n100 -#### A masked pattern was here #### PREHOOK: query: create table cmv_basetable_2_n100 (a int, b varchar(256), c decimal(10,2), d int) stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default @@ -62,16 +52,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n100.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n100.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n100.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n100.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n100 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n100 -PREHOOK: Output: default@cmv_basetable_2_n100 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n100 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n100 -POSTHOOK: Output: default@cmv_basetable_2_n100 -#### A masked pattern was here #### PREHOOK: query: CREATE MATERIALIZED VIEW cmv_mat_view_n300 AS SELECT cmv_basetable_n100.a, cmv_basetable_2_n100.c FROM cmv_basetable_n100 JOIN cmv_basetable_2_n100 ON (cmv_basetable_n100.a = cmv_basetable_2_n100.a) @@ -109,7 +89,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 numFiles 2 numRows 2 @@ -151,16 +131,6 @@ POSTHOOK: Lineage: cmv_basetable_2_n100.a SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n100.b SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n100.c SCRIPT [] POSTHOOK: Lineage: cmv_basetable_2_n100.d SCRIPT [] -PREHOOK: query: analyze table cmv_basetable_2_n100 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@cmv_basetable_2_n100 -PREHOOK: Output: default@cmv_basetable_2_n100 -#### A masked pattern was here #### -POSTHOOK: query: analyze table cmv_basetable_2_n100 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@cmv_basetable_2_n100 -POSTHOOK: Output: default@cmv_basetable_2_n100 -#### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED cmv_mat_view_n300 PREHOOK: type: DESCTABLE PREHOOK: Input: default@cmv_mat_view_n300 @@ -178,7 +148,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 numFiles 2 numRows 2 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_describe.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_describe.q.out index 581cc87d26..00577c052c 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_describe.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_describe.q.out @@ -66,7 +66,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"c\":\"true\"}} bucketing_version 2 comment this is the first view key foo @@ -94,7 +94,7 @@ PREHOOK: query: show tblproperties cmv_mat_view_n8 PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: show tblproperties cmv_mat_view_n8 POSTHOOK: type: SHOW_TBLPROPERTIES -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","c":"true"}} bucketing_version 2 comment this is the first view key foo @@ -157,7 +157,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\"}} bucketing_version 2 comment this is the second view key alice @@ -239,7 +239,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\",\"b\":\"true\",\"c\":\"true\"}} bucketing_version 2 comment this is the third view numFiles 1 @@ -330,7 +330,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MATERIALIZED_VIEW Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\"}} bucketing_version 2 comment this is the last view numFiles 1 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_partitioned.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_partitioned.q.out index ba4d76f89f..2c9b36d803 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_partitioned.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_partitioned.q.out @@ -38,6 +38,9 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) #### A masked pattern was here #### Vertices: Map 1 @@ -45,24 +48,60 @@ STAGE PLANS: TableScan alias: src_txn filterExpr: ((UDFToDouble(key) > 200.0D) and (UDFToDouble(key) < 250.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 175168 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToDouble(key) < 250.0D) and (UDFToDouble(key) > 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 19268 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: value (type: string), key (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 19268 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 55 Data size: 19268 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.partition_mv_1 - Execution mode: vectorized, llap + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: col1, col2 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll') + keys: col2 (type: string) + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct) + Execution mode: llap LLAP IO: may be used (ACID table) + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-2 Dependency Collection @@ -92,6 +131,10 @@ STAGE PLANS: Stage: Stage-3 Stats Work Basic Stats Work: + Column Stats Desc: + Columns: value + Column Types: string + Table: default.partition_mv_1 Stage: Stage-5 Materialized View Work @@ -578,17 +621,17 @@ STAGE PLANS: TableScan alias: src_txn filterExpr: ((UDFToDouble(key) > 200.0D) and (UDFToDouble(key) < 250.0D)) (type: boolean) - Statistics: Num rows: 501 Data size: 175536 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 501 Data size: 90180 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToDouble(key) < 250.0D) and (UDFToDouble(key) > 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 19270 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9900 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: value (type: string), key (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 19270 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9900 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 55 Data size: 19270 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9900 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat @@ -597,18 +640,18 @@ STAGE PLANS: Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: value, key - Statistics: Num rows: 55 Data size: 19270 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9900 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: compute_stats(value, 'hll') keys: key (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 19270 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 19270 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: struct) Execution mode: llap LLAP IO: may be used (ACID table) @@ -620,14 +663,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 9459 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: struct), _col0 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 9459 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 27 Data size: 9459 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -956,19 +999,19 @@ STAGE PLANS: TableScan alias: src_txn filterExpr: ((UDFToDouble(key) > 200.0D) and (UDFToDouble(key) < 250.0D)) (type: boolean) - Statistics: Num rows: 502 Data size: 175904 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 502 Data size: 90862 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToDouble(key) < 250.0D) and (UDFToDouble(key) > 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 19272 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 19272 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 19272 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: string) Execution mode: vectorized, llap LLAP IO: may be used (ACID table) @@ -977,19 +1020,19 @@ STAGE PLANS: TableScan alias: src_txn_2 filterExpr: ((UDFToDouble(key) > 200.0D) and (UDFToDouble(key) < 250.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 87584 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToDouble(key) < 250.0D) and (UDFToDouble(key) > 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9634 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 4785 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 55 Data size: 9634 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 4785 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9634 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 4785 Basic stats: COMPLETE Column stats: COMPLETE Execution mode: vectorized, llap LLAP IO: may be used (ACID table) Reducer 2 @@ -1002,14 +1045,14 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 60 Data size: 21199 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col0 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 60 Data size: 21199 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 60 Data size: 21199 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat @@ -1018,18 +1061,18 @@ STAGE PLANS: Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: value, key - Statistics: Num rows: 60 Data size: 21199 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9955 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: compute_stats(value, 'hll') keys: key (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 60 Data size: 21199 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 60 Data size: 21199 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: struct) Reducer 3 Execution mode: llap @@ -1039,14 +1082,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 30 Data size: 10599 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: struct), _col0 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 30 Data size: 10599 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 30 Data size: 10599 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 14229 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_partitioned_3.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_partitioned_3.q.out index 4ed30ee8a0..e7b3e7ca66 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_partitioned_3.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_partitioned_3.q.out @@ -48,14 +48,14 @@ STAGE PLANS: TableScan alias: src_txn filterExpr: ((UDFToDouble(key) > 200.0D) and (UDFToDouble(key) < 250.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 175168 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToDouble(key) < 250.0D) and (UDFToDouble(key) > 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 19268 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: value (type: string), key (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 19268 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: string) sort order: + @@ -72,7 +72,7 @@ STAGE PLANS: File Output Operator compressed: false Dp Sort State: PARTITION_SORTED - Statistics: Num rows: 55 Data size: 19268 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat @@ -107,6 +107,10 @@ STAGE PLANS: Stage: Stage-3 Stats Work Basic Stats Work: + Column Stats Desc: + Columns: value + Column Types: string + Table: default.partition_mv_sdp Stage: Stage-5 Materialized View Work diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_1.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_1.q.out index 735a4db2ee..5714ef80ca 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_1.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_1.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n3.deptno SCRIPT [] POSTHOOK: Lineage: emps_n3.empid SCRIPT [] POSTHOOK: Lineage: emps_n3.name SCRIPT [] POSTHOOK: Lineage: emps_n3.salary SCRIPT [] -PREHOOK: query: analyze table emps_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n3 -PREHOOK: Output: default@emps_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n3 -POSTHOOK: Output: default@emps_n3 -#### A masked pattern was here #### PREHOOK: query: create table depts_n2 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n2 POSTHOOK: Lineage: depts_n2.deptno SCRIPT [] POSTHOOK: Lineage: depts_n2.locationid SCRIPT [] POSTHOOK: Lineage: depts_n2.name SCRIPT [] -PREHOOK: query: analyze table depts_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n2 -PREHOOK: Output: default@depts_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n2 -POSTHOOK: Output: default@depts_n2 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n2 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n2 POSTHOOK: Lineage: dependents_n2.empid SCRIPT [] POSTHOOK: Lineage: dependents_n2.name SCRIPT [] -PREHOOK: query: analyze table dependents_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n2 -PREHOOK: Output: default@dependents_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n2 -POSTHOOK: Output: default@dependents_n2 -#### A masked pattern was here #### PREHOOK: query: create table locations_n2 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n2 POSTHOOK: Lineage: locations_n2.locationid SCRIPT [] POSTHOOK: Lineage: locations_n2.name SCRIPT [] -PREHOOK: query: analyze table locations_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n2 -PREHOOK: Output: default@locations_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n2 -POSTHOOK: Output: default@locations_n2 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n3 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n3 add constraint pk1 primary key (empid) disable novalidate rely @@ -184,16 +144,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select * from (select * from emps_n3 where empid < 120) t @@ -331,16 +281,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n3.name, emps_n3.salary, emps_n3.commission from emps_n3 @@ -422,16 +362,6 @@ POSTHOOK: Input: default@depts_n2 POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select empid deptno from emps_n3 join depts_n2 using (deptno) where empid = 1 @@ -500,16 +430,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select * from emps_n3 where empid > 120 union all select * from emps_n3 where empid < 150 @@ -627,16 +547,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select empid, deptno from emps_n3 group by empid, deptno PREHOOK: type: QUERY @@ -695,16 +605,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select empid, name from emps_n3 group by empid, name PREHOOK: type: QUERY @@ -763,16 +663,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select name, salary from emps_n3 group by name, salary PREHOOK: type: QUERY @@ -834,16 +724,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n3 group by name PREHOOK: type: QUERY @@ -945,16 +825,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n3 where deptno = 10 group by name PREHOOK: type: QUERY @@ -1057,16 +927,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n3 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n2 -PREHOOK: query: analyze table mv1_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n2 -PREHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n2 -POSTHOOK: Output: default@mv1_n2 -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n3 group by name PREHOOK: type: QUERY diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_10.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_10.q.out index 3ed5609396..94940ac572 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_10.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_10.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n10.deptno SCRIPT [] POSTHOOK: Lineage: emps_n10.empid SCRIPT [] POSTHOOK: Lineage: emps_n10.name SCRIPT [] POSTHOOK: Lineage: emps_n10.salary SCRIPT [] -PREHOOK: query: analyze table emps_n10 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n10 -PREHOOK: Output: default@emps_n10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n10 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n10 -POSTHOOK: Output: default@emps_n10 -#### A masked pattern was here #### PREHOOK: query: create materialized view mv1_n10 as select deptno, sum(salary), count(salary) as a from emps_n10 group by deptno @@ -57,16 +47,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n10 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n10 -PREHOOK: query: analyze table mv1_n10 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n10 -PREHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n10 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n10 -POSTHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### PREHOOK: query: explain select deptno, avg(salary) as a from emps_n10 group by deptno @@ -132,16 +112,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n10 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n10 -PREHOOK: query: analyze table mv1_n10 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n10 -PREHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n10 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n10 -POSTHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### PREHOOK: query: explain select salary, avg(salary) as a from emps_n10 group by salary @@ -208,16 +178,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n10 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n10 -PREHOOK: query: analyze table mv1_n10 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n10 -PREHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n10 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n10 -POSTHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### PREHOOK: query: explain select salary, avg(salary) as a from emps_n10 where salary > 0 group by salary @@ -305,16 +265,6 @@ POSTHOOK: Lineage: emps_n10_2.deptno SCRIPT [] POSTHOOK: Lineage: emps_n10_2.empid SCRIPT [] POSTHOOK: Lineage: emps_n10_2.name SCRIPT [] POSTHOOK: Lineage: emps_n10_2.salary SCRIPT [] -PREHOOK: query: analyze table emps_n10_2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n10_2 -PREHOOK: Output: default@emps_n10_2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n10_2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n10_2 -POSTHOOK: Output: default@emps_n10_2 -#### A masked pattern was here #### PREHOOK: query: create materialized view mv1_n10 as select salary, sum(salary), count(salary) as a from emps_n10_2 where salary > 0 group by salary @@ -329,16 +279,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n10_2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n10 -PREHOOK: query: analyze table mv1_n10 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n10 -PREHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n10 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n10 -POSTHOOK: Output: default@mv1_n10 -#### A masked pattern was here #### PREHOOK: query: explain select avg(salary) from emps_n10_2 where salary > 0 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_2.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_2.q.out index d5afe3fc8f..f87b55b5c2 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_2.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_2.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n0.deptno SCRIPT [] POSTHOOK: Lineage: emps_n0.empid SCRIPT [] POSTHOOK: Lineage: emps_n0.name SCRIPT [] POSTHOOK: Lineage: emps_n0.salary SCRIPT [] -PREHOOK: query: analyze table emps_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n0 -PREHOOK: Output: default@emps_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n0 -POSTHOOK: Output: default@emps_n0 -#### A masked pattern was here #### PREHOOK: query: create table depts_n0 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n0 POSTHOOK: Lineage: depts_n0.deptno SCRIPT [] POSTHOOK: Lineage: depts_n0.locationid SCRIPT [] POSTHOOK: Lineage: depts_n0.name SCRIPT [] -PREHOOK: query: analyze table depts_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n0 -PREHOOK: Output: default@depts_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n0 -POSTHOOK: Output: default@depts_n0 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n0 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n0 POSTHOOK: Lineage: dependents_n0.empid SCRIPT [] POSTHOOK: Lineage: dependents_n0.name SCRIPT [] -PREHOOK: query: analyze table dependents_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n0 -PREHOOK: Output: default@dependents_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n0 -POSTHOOK: Output: default@dependents_n0 -#### A masked pattern was here #### PREHOOK: query: create table locations_n0 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n0 POSTHOOK: Lineage: locations_n0.locationid SCRIPT [] POSTHOOK: Lineage: locations_n0.name SCRIPT [] -PREHOOK: query: analyze table locations_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n0 -PREHOOK: Output: default@locations_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n0 -POSTHOOK: Output: default@locations_n0 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n0 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n0 add constraint pk1 primary key (empid) disable novalidate rely @@ -190,16 +150,6 @@ POSTHOOK: Input: default@depts_n0 POSTHOOK: Input: default@emps_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n0 -PREHOOK: query: analyze table mv1_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n0 -PREHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n0 -POSTHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n0 join depts_n0 using (deptno) where depts_n0.deptno > 20 @@ -278,16 +228,6 @@ POSTHOOK: Input: default@depts_n0 POSTHOOK: Input: default@emps_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n0 -PREHOOK: query: analyze table mv1_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n0 -PREHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n0 -POSTHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n0 join depts_n0 using (deptno) where depts_n0.deptno > 20 @@ -366,16 +306,6 @@ POSTHOOK: Input: default@depts_n0 POSTHOOK: Input: default@emps_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n0 -PREHOOK: query: analyze table mv1_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n0 -PREHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n0 -POSTHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n0 join depts_n0 using (deptno) where depts_n0.deptno > 20 @@ -454,16 +384,6 @@ POSTHOOK: Input: default@depts_n0 POSTHOOK: Input: default@emps_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n0 -PREHOOK: query: analyze table mv1_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n0 -PREHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n0 -POSTHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n0.deptno from depts_n0 join emps_n0 using (deptno) where emps_n0.empid > 15 @@ -547,16 +467,6 @@ POSTHOOK: Input: default@depts_n0 POSTHOOK: Input: default@emps_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n0 -PREHOOK: query: analyze table mv1_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n0 -PREHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n0 -POSTHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n0.deptno from depts_n0 join emps_n0 using (deptno) where emps_n0.empid > 15 @@ -693,7 +603,7 @@ POSTHOOK: query: drop materialized view mv1_n0 POSTHOOK: type: DROP_MATERIALIZED_VIEW POSTHOOK: Input: default@mv1_n0 POSTHOOK: Output: default@mv1_n0 -Warning: Shuffle Join MERGEJOIN[18][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[28][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 2' is a cross product PREHOOK: query: create materialized view mv1_n0 as select depts_n0.name, dependents_n0.name as name2, emps_n0.deptno, depts_n0.deptno as deptno2, dependents_n0.empid from depts_n0, dependents_n0, emps_n0 @@ -716,16 +626,6 @@ POSTHOOK: Input: default@depts_n0 POSTHOOK: Input: default@emps_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n0 -PREHOOK: query: analyze table mv1_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n0 -PREHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n0 -POSTHOOK: Output: default@mv1_n0 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n0.empid from depts_n0 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_3.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_3.q.out index 91f6982e19..93bda7650b 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_3.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_3.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n9.deptno SCRIPT [] POSTHOOK: Lineage: emps_n9.empid SCRIPT [] POSTHOOK: Lineage: emps_n9.name SCRIPT [] POSTHOOK: Lineage: emps_n9.salary SCRIPT [] -PREHOOK: query: analyze table emps_n9 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n9 -PREHOOK: Output: default@emps_n9 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n9 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n9 -POSTHOOK: Output: default@emps_n9 -#### A masked pattern was here #### PREHOOK: query: create table depts_n7 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n7 POSTHOOK: Lineage: depts_n7.deptno SCRIPT [] POSTHOOK: Lineage: depts_n7.locationid SCRIPT [] POSTHOOK: Lineage: depts_n7.name SCRIPT [] -PREHOOK: query: analyze table depts_n7 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n7 -PREHOOK: Output: default@depts_n7 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n7 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n7 -POSTHOOK: Output: default@depts_n7 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n5 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n5 POSTHOOK: Lineage: dependents_n5.empid SCRIPT [] POSTHOOK: Lineage: dependents_n5.name SCRIPT [] -PREHOOK: query: analyze table dependents_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n5 -PREHOOK: Output: default@dependents_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n5 -POSTHOOK: Output: default@dependents_n5 -#### A masked pattern was here #### PREHOOK: query: create table locations_n5 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n5 POSTHOOK: Lineage: locations_n5.locationid SCRIPT [] POSTHOOK: Lineage: locations_n5.name SCRIPT [] -PREHOOK: query: analyze table locations_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n5 -PREHOOK: Output: default@locations_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n5 -POSTHOOK: Output: default@locations_n5 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n9 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n9 add constraint pk1 primary key (empid) disable novalidate rely @@ -188,16 +148,6 @@ POSTHOOK: Input: default@depts_n7 POSTHOOK: Input: default@emps_n9 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n5 -PREHOOK: query: analyze table mv1_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n5 -PREHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n5 -POSTHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### PREHOOK: query: explain select empid deptno from emps_n9 join depts_n7 using (deptno) where empid = 1 @@ -270,16 +220,6 @@ POSTHOOK: Input: default@depts_n7 POSTHOOK: Input: default@emps_n9 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n5 -PREHOOK: query: analyze table mv1_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n5 -PREHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n5 -POSTHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### PREHOOK: query: explain select empid deptno from emps_n9 join depts_n7 using (deptno) where empid > 1 @@ -356,16 +296,6 @@ POSTHOOK: Input: default@depts_n7 POSTHOOK: Input: default@emps_n9 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n5 -PREHOOK: query: analyze table mv1_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n5 -PREHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n5 -POSTHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### PREHOOK: query: explain select empid deptno from emps_n9 join depts_n7 using (deptno) where empid = 1 @@ -440,16 +370,6 @@ POSTHOOK: Input: default@depts_n7 POSTHOOK: Input: default@emps_n9 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n5 -PREHOOK: query: analyze table mv1_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n5 -PREHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n5 -POSTHOOK: Output: default@mv1_n5 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n5.empid from depts_n7 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_4.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_4.q.out index 18efcba19f..2442ecc0df 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_4.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_4.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n5.deptno SCRIPT [] POSTHOOK: Lineage: emps_n5.empid SCRIPT [] POSTHOOK: Lineage: emps_n5.name SCRIPT [] POSTHOOK: Lineage: emps_n5.salary SCRIPT [] -PREHOOK: query: analyze table emps_n5 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n5 -PREHOOK: Output: default@emps_n5 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n5 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n5 -POSTHOOK: Output: default@emps_n5 -#### A masked pattern was here #### PREHOOK: query: create table depts_n4 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n4 POSTHOOK: Lineage: depts_n4.deptno SCRIPT [] POSTHOOK: Lineage: depts_n4.locationid SCRIPT [] POSTHOOK: Lineage: depts_n4.name SCRIPT [] -PREHOOK: query: analyze table depts_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n4 -PREHOOK: Output: default@depts_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n4 -POSTHOOK: Output: default@depts_n4 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n3 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n3 POSTHOOK: Lineage: dependents_n3.empid SCRIPT [] POSTHOOK: Lineage: dependents_n3.name SCRIPT [] -PREHOOK: query: analyze table dependents_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n3 -PREHOOK: Output: default@dependents_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n3 -POSTHOOK: Output: default@dependents_n3 -#### A masked pattern was here #### PREHOOK: query: create table locations_n3 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n3 POSTHOOK: Lineage: locations_n3.locationid SCRIPT [] POSTHOOK: Lineage: locations_n3.name SCRIPT [] -PREHOOK: query: analyze table locations_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n3 -PREHOOK: Output: default@locations_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n3 -POSTHOOK: Output: default@locations_n3 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n5 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n5 add constraint pk1 primary key (empid) disable novalidate rely @@ -186,16 +146,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select name, count(*) as c, sum(empid) as s from emps_n5 group by name @@ -310,16 +260,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select salary, name, sum(empid) as s, count(*) as c from emps_n5 group by name, salary @@ -391,16 +331,6 @@ POSTHOOK: Input: default@depts_n4 POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n4.deptno, count(*) as c, sum(empid) as s from emps_n5 join depts_n4 using (deptno) @@ -525,16 +455,6 @@ POSTHOOK: Input: default@depts_n4 POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n4.deptno, sum(empid) as s from emps_n5 join depts_n4 using (deptno) @@ -654,16 +574,6 @@ POSTHOOK: Input: default@depts_n4 POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n4.deptno, sum(empid) + 1 as s from emps_n5 join depts_n4 using (deptno) @@ -793,16 +703,6 @@ POSTHOOK: Input: default@depts_n4 POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n3.empid, sum(salary) as s from emps_n5 @@ -978,16 +878,6 @@ POSTHOOK: Input: default@dependents_n3 POSTHOOK: Input: default@emps_n5 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n3 -PREHOOK: query: analyze table mv1_n3 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n3 -PREHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n3 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n3 -POSTHOOK: Output: default@mv1_n3 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n5.deptno, count(distinct salary) as s from emps_n5 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_5.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_5.q.out index e4c939e8bf..b865675e8a 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_5.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_5.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n2.deptno SCRIPT [] POSTHOOK: Lineage: emps_n2.empid SCRIPT [] POSTHOOK: Lineage: emps_n2.name SCRIPT [] POSTHOOK: Lineage: emps_n2.salary SCRIPT [] -PREHOOK: query: analyze table emps_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n2 -PREHOOK: Output: default@emps_n2 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n2 -POSTHOOK: Output: default@emps_n2 -#### A masked pattern was here #### PREHOOK: query: create table depts_n1 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n1 POSTHOOK: Lineage: depts_n1.deptno SCRIPT [] POSTHOOK: Lineage: depts_n1.locationid SCRIPT [] POSTHOOK: Lineage: depts_n1.name SCRIPT [] -PREHOOK: query: analyze table depts_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n1 -PREHOOK: Output: default@depts_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n1 -POSTHOOK: Output: default@depts_n1 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n1 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n1 POSTHOOK: Lineage: dependents_n1.empid SCRIPT [] POSTHOOK: Lineage: dependents_n1.name SCRIPT [] -PREHOOK: query: analyze table dependents_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n1 -PREHOOK: Output: default@dependents_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n1 -POSTHOOK: Output: default@dependents_n1 -#### A masked pattern was here #### PREHOOK: query: create table locations_n1 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n1 POSTHOOK: Lineage: locations_n1.locationid SCRIPT [] POSTHOOK: Lineage: locations_n1.name SCRIPT [] -PREHOOK: query: analyze table locations_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n1 -PREHOOK: Output: default@locations_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n1 -POSTHOOK: Output: default@locations_n1 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n2 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n2 add constraint pk1 primary key (empid) disable novalidate rely @@ -200,16 +160,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n2 where deptno >= 20 group by name PREHOOK: type: QUERY @@ -314,16 +264,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select name, sum(empid) as s from emps_n2 where deptno > 15 group by name @@ -451,16 +391,6 @@ POSTHOOK: Input: default@emps_n2 POSTHOOK: Input: default@locations_n1 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n1.empid from depts_n1 @@ -604,16 +534,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select deptno from emps_n2 group by deptno PREHOOK: type: QUERY @@ -719,16 +639,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select deptno, empid, sum(empid) as s, count(*) as c from emps_n2 group by empid, deptno @@ -803,16 +713,6 @@ POSTHOOK: Input: default@dependents_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n1.empid, sum(salary) as s from emps_n2 @@ -945,16 +845,6 @@ POSTHOOK: Input: default@dependents_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n1.name, sum(salary) as s from emps_n2 @@ -1102,7 +992,7 @@ POSTHOOK: query: drop materialized view mv1_n1 POSTHOOK: type: DROP_MATERIALIZED_VIEW POSTHOOK: Input: default@mv1_n1 POSTHOOK: Output: default@mv1_n1 -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[23][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: create materialized view mv1_n1 as select a.empid deptno from (select * from emps_n2 where empid = 1) a @@ -1125,16 +1015,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select a.empid from (select * from emps_n2 where empid = 1) a @@ -1192,7 +1072,7 @@ POSTHOOK: query: drop materialized view mv1_n1 POSTHOOK: type: DROP_MATERIALIZED_VIEW POSTHOOK: Input: default@mv1_n1 POSTHOOK: Output: default@mv1_n1 -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[23][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: create materialized view mv1_n1 as select a.empid, a.deptno from (select * from emps_n2 where empid = 1) a @@ -1215,16 +1095,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select a.empid from (select * from emps_n2 where empid = 1) a @@ -1300,16 +1170,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n2 where empid = 1 PREHOOK: type: QUERY @@ -1355,7 +1215,7 @@ POSTHOOK: query: drop materialized view mv1_n1 POSTHOOK: type: DROP_MATERIALIZED_VIEW POSTHOOK: Input: default@mv1_n1 POSTHOOK: Output: default@mv1_n1 -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[23][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: create materialized view mv1_n1 as select emps_n2.empid, emps_n2.deptno from emps_n2 join depts_n1 on (emps_n2.deptno = depts_n1.deptno) @@ -1378,16 +1238,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n2.empid from emps_n2 join dependents_n1 on (emps_n2.empid = dependents_n1.empid) @@ -1445,7 +1295,7 @@ POSTHOOK: query: drop materialized view mv1_n1 POSTHOOK: type: DROP_MATERIALIZED_VIEW POSTHOOK: Input: default@mv1_n1 POSTHOOK: Output: default@mv1_n1 -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[23][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: create materialized view mv1_n1 as select emps_n2.empid, emps_n2.deptno from emps_n2 join depts_n1 a on (emps_n2.deptno=a.deptno) @@ -1470,16 +1320,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n2.empid from emps_n2 join dependents_n1 on (emps_n2.empid = dependents_n1.empid) @@ -1561,16 +1401,6 @@ POSTHOOK: Input: default@depts_n1 POSTHOOK: Input: default@emps_n2 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n1 -PREHOOK: query: analyze table mv1_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n1 -PREHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n1 -POSTHOOK: Output: default@mv1_n1 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n2.empid from emps_n2 join dependents_n1 on (emps_n2.empid = dependents_n1.empid) diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_6.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_6.q.out index 899e50a4ed..b66cd57acb 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_6.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_6.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps.deptno SCRIPT [] POSTHOOK: Lineage: emps.empid SCRIPT [] POSTHOOK: Lineage: emps.name SCRIPT [] POSTHOOK: Lineage: emps.salary SCRIPT [] -PREHOOK: query: analyze table emps compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps -PREHOOK: Output: default@emps -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps -POSTHOOK: Output: default@emps -#### A masked pattern was here #### PREHOOK: query: create table depts ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts POSTHOOK: Lineage: depts.deptno SCRIPT [] POSTHOOK: Lineage: depts.locationid SCRIPT [] POSTHOOK: Lineage: depts.name SCRIPT [] -PREHOOK: query: analyze table depts compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts -PREHOOK: Output: default@depts -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts -POSTHOOK: Output: default@depts -#### A masked pattern was here #### PREHOOK: query: create table dependents ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents POSTHOOK: Lineage: dependents.empid SCRIPT [] POSTHOOK: Lineage: dependents.name SCRIPT [] -PREHOOK: query: analyze table dependents compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents -PREHOOK: Output: default@dependents -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents -POSTHOOK: Output: default@dependents -#### A masked pattern was here #### PREHOOK: query: create table locations ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations POSTHOOK: Lineage: locations.locationid SCRIPT [] POSTHOOK: Lineage: locations.name SCRIPT [] -PREHOOK: query: analyze table locations compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations -PREHOOK: Output: default@locations -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations -POSTHOOK: Output: default@locations -#### A masked pattern was here #### PREHOOK: query: alter table emps add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps add constraint pk1 primary key (empid) disable novalidate rely @@ -202,16 +162,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1 -PREHOOK: query: analyze table mv1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1 -PREHOOK: Output: default@mv1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1 -POSTHOOK: Output: default@mv1 -#### A masked pattern was here #### PREHOOK: query: explain select salary, sum(empid) + 1 as s from emps where deptno > 10 group by salary @@ -327,16 +277,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1 -PREHOOK: query: analyze table mv1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1 -PREHOOK: Output: default@mv1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1 -POSTHOOK: Output: default@mv1 -#### A masked pattern was here #### PREHOOK: query: explain select salary + 1, sum(empid) + 1 as s from emps where deptno > 15 group by salary @@ -456,16 +396,6 @@ POSTHOOK: Input: default@depts POSTHOOK: Input: default@emps POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1 -PREHOOK: query: analyze table mv1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1 -PREHOOK: Output: default@mv1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1 -POSTHOOK: Output: default@mv1 -#### A masked pattern was here #### PREHOOK: query: explain select dependents.empid from emps @@ -616,16 +546,6 @@ POSTHOOK: Input: default@depts POSTHOOK: Input: default@emps POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1 -PREHOOK: query: analyze table mv1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1 -PREHOOK: Output: default@mv1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1 -POSTHOOK: Output: default@mv1 -#### A masked pattern was here #### PREHOOK: query: explain select dependents.empid from depts @@ -804,16 +724,6 @@ POSTHOOK: Input: default@dependents POSTHOOK: Input: default@emps POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1 -PREHOOK: query: analyze table mv1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1 -PREHOOK: Output: default@mv1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1 -POSTHOOK: Output: default@mv1 -#### A masked pattern was here #### PREHOOK: query: explain select emps.empid, dependents.empid, emps.deptno from emps diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_7.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_7.q.out index 902a8d450f..03acafaca8 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_7.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_7.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n8.deptno SCRIPT [] POSTHOOK: Lineage: emps_n8.empid SCRIPT [] POSTHOOK: Lineage: emps_n8.name SCRIPT [] POSTHOOK: Lineage: emps_n8.salary SCRIPT [] -PREHOOK: query: analyze table emps_n8 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n8 -PREHOOK: Output: default@emps_n8 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n8 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n8 -POSTHOOK: Output: default@emps_n8 -#### A masked pattern was here #### PREHOOK: query: create table depts_n6 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n6 POSTHOOK: Lineage: depts_n6.deptno SCRIPT [] POSTHOOK: Lineage: depts_n6.locationid SCRIPT [] POSTHOOK: Lineage: depts_n6.name SCRIPT [] -PREHOOK: query: analyze table depts_n6 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n6 -PREHOOK: Output: default@depts_n6 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n6 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n6 -POSTHOOK: Output: default@depts_n6 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n4 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n4 POSTHOOK: Lineage: dependents_n4.empid SCRIPT [] POSTHOOK: Lineage: dependents_n4.name SCRIPT [] -PREHOOK: query: analyze table dependents_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n4 -PREHOOK: Output: default@dependents_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n4 -POSTHOOK: Output: default@dependents_n4 -#### A masked pattern was here #### PREHOOK: query: create table locations_n4 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n4 POSTHOOK: Lineage: locations_n4.locationid SCRIPT [] POSTHOOK: Lineage: locations_n4.name SCRIPT [] -PREHOOK: query: analyze table locations_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n4 -PREHOOK: Output: default@locations_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n4 -POSTHOOK: Output: default@locations_n4 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n8 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n8 add constraint pk1 primary key (empid) disable novalidate rely @@ -218,16 +178,6 @@ POSTHOOK: Input: default@emps_n8 POSTHOOK: Input: default@locations_n4 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n4 -PREHOOK: query: analyze table mv1_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n4 -PREHOOK: Output: default@mv1_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n4 -POSTHOOK: Output: default@mv1_n4 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n4.empid, depts_n6.deptno from depts_n6 @@ -525,16 +475,6 @@ POSTHOOK: Input: default@emps_n8 POSTHOOK: Input: default@locations_n4 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n4 -PREHOOK: query: analyze table mv1_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n4 -PREHOOK: Output: default@mv1_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n4 -POSTHOOK: Output: default@mv1_n4 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n4.empid, count(emps_n8.salary) + 1 from depts_n6 @@ -864,16 +804,6 @@ POSTHOOK: Input: default@depts_n6 POSTHOOK: Input: default@emps_n8 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n4 -PREHOOK: query: analyze table mv1_n4 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n4 -PREHOOK: Output: default@mv1_n4 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n4 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n4 -POSTHOOK: Output: default@mv1_n4 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n4.empid from depts_n6 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_8.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_8.q.out index aa8763f470..a11c846864 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_8.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_8.q.out @@ -40,16 +40,6 @@ POSTHOOK: Lineage: source_table_001.my_date SCRIPT [] POSTHOOK: Lineage: source_table_001.my_id SCRIPT [] POSTHOOK: Lineage: source_table_001.my_id2 SCRIPT [] POSTHOOK: Lineage: source_table_001.up_volume SCRIPT [] -PREHOOK: query: analyze table source_table_001 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@source_table_001 -PREHOOK: Output: default@source_table_001 -#### A masked pattern was here #### -POSTHOOK: query: analyze table source_table_001 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@source_table_001 -POSTHOOK: Output: default@source_table_001 -#### A masked pattern was here #### PREHOOK: query: CREATE MATERIALIZED VIEW source_table_001_mv AS SELECT SUM(A.DOWN_VOLUME) AS DOWN_VOLUME_SUM, @@ -72,16 +62,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@source_table_001 POSTHOOK: Output: database:default POSTHOOK: Output: default@source_table_001_mv -PREHOOK: query: analyze table source_table_001_mv compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@source_table_001_mv -PREHOOK: Output: default@source_table_001_mv -#### A masked pattern was here #### -POSTHOOK: query: analyze table source_table_001_mv compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@source_table_001_mv -POSTHOOK: Output: default@source_table_001_mv -#### A masked pattern was here #### PREHOOK: query: explain select SUM(A.DOWN_VOLUME) AS DOWNLOAD_VOLUME_BYTES, diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_9.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_9.q.out index 3c1de08e2b..ea504950fc 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_9.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_9.q.out @@ -40,16 +40,6 @@ POSTHOOK: Lineage: source_table_001_n0.my_date SCRIPT [] POSTHOOK: Lineage: source_table_001_n0.my_id SCRIPT [] POSTHOOK: Lineage: source_table_001_n0.my_id2 SCRIPT [] POSTHOOK: Lineage: source_table_001_n0.up_volume SCRIPT [] -PREHOOK: query: analyze table source_table_001_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@source_table_001_n0 -PREHOOK: Output: default@source_table_001_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table source_table_001_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@source_table_001_n0 -POSTHOOK: Output: default@source_table_001_n0 -#### A masked pattern was here #### PREHOOK: query: CREATE MATERIALIZED VIEW source_table_001_mv_n0 AS SELECT SUM(A.DOWN_VOLUME) AS DOWN_VOLUME_SUM, @@ -72,16 +62,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@source_table_001_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@source_table_001_mv_n0 -PREHOOK: query: analyze table source_table_001_mv_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@source_table_001_mv_n0 -PREHOOK: Output: default@source_table_001_mv_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table source_table_001_mv_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@source_table_001_mv_n0 -POSTHOOK: Output: default@source_table_001_mv_n0 -#### A masked pattern was here #### PREHOOK: query: explain select SUM(A.DOWN_VOLUME) AS DOWNLOAD_VOLUME_BYTES, @@ -194,16 +174,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@source_table_001_n0 POSTHOOK: Output: database:default POSTHOOK: Output: default@source_table_001_mv_n0 -PREHOOK: query: analyze table source_table_001_mv_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@source_table_001_mv_n0 -PREHOOK: Output: default@source_table_001_mv_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table source_table_001_mv_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@source_table_001_mv_n0 -POSTHOOK: Output: default@source_table_001_mv_n0 -#### A masked pattern was here #### PREHOOK: query: explain select SUM(A.DOWN_VOLUME) AS DOWNLOAD_VOLUME_BYTES, diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_empty.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_empty.q.out index 610ab6ddbe..61a93651b1 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_empty.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_empty.q.out @@ -18,16 +18,6 @@ stored as orc TBLPROPERTIES ('transactional'='true') POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@emps_mv_rewrite_empty -PREHOOK: query: analyze table emps_mv_rewrite_empty compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_mv_rewrite_empty -PREHOOK: Output: default@emps_mv_rewrite_empty -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_mv_rewrite_empty compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_mv_rewrite_empty -POSTHOOK: Output: default@emps_mv_rewrite_empty -#### A masked pattern was here #### PREHOOK: query: create materialized view emps_mv_rewrite_empty_mv1 as select * from emps_mv_rewrite_empty where empid < 150 PREHOOK: type: CREATE_MATERIALIZED_VIEW diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt.q.out index 806df43721..b022ee8ff5 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n30.deptno SCRIPT [] POSTHOOK: Lineage: emps_n30.empid SCRIPT [] POSTHOOK: Lineage: emps_n30.name SCRIPT [] POSTHOOK: Lineage: emps_n30.salary SCRIPT [] -PREHOOK: query: analyze table emps_n30 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n30 -PREHOOK: Output: default@emps_n30 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n30 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n30 -POSTHOOK: Output: default@emps_n30 -#### A masked pattern was here #### PREHOOK: query: create table depts_n20 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n20 POSTHOOK: Lineage: depts_n20.deptno SCRIPT [] POSTHOOK: Lineage: depts_n20.locationid SCRIPT [] POSTHOOK: Lineage: depts_n20.name SCRIPT [] -PREHOOK: query: analyze table depts_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n20 -PREHOOK: Output: default@depts_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n20 -POSTHOOK: Output: default@depts_n20 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n20 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n20 POSTHOOK: Lineage: dependents_n20.empid SCRIPT [] POSTHOOK: Lineage: dependents_n20.name SCRIPT [] -PREHOOK: query: analyze table dependents_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n20 -PREHOOK: Output: default@dependents_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n20 -POSTHOOK: Output: default@dependents_n20 -#### A masked pattern was here #### PREHOOK: query: create table locations_n20 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n20 POSTHOOK: Lineage: locations_n20.locationid SCRIPT [] POSTHOOK: Lineage: locations_n20.name SCRIPT [] -PREHOOK: query: analyze table locations_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n20 -PREHOOK: Output: default@locations_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n20 -POSTHOOK: Output: default@locations_n20 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely @@ -186,16 +146,6 @@ POSTHOOK: type: CREATE_MATERIALIZED_VIEW POSTHOOK: Input: default@emps_n30 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n20 -PREHOOK: query: analyze table mv1_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n20 -PREHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n20 -POSTHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n30.name, emps_n30.salary, emps_n30.commission from emps_n30 @@ -339,16 +289,6 @@ POSTHOOK: Input: default@depts_n20 POSTHOOK: Input: default@emps_n30 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n20 -PREHOOK: query: analyze table mv1_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n20 -PREHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n20 -POSTHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n20.deptno, count(*) as c, sum(empid) as s from emps_n30 join depts_n20 using (deptno) @@ -475,16 +415,6 @@ POSTHOOK: Input: default@dependents_n20 POSTHOOK: Input: default@emps_n30 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n20 -PREHOOK: query: analyze table mv1_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n20 -PREHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n20 -POSTHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n20.empid, sum(salary) as s from emps_n30 @@ -651,16 +581,6 @@ POSTHOOK: Input: default@dependents_n20 POSTHOOK: Input: default@emps_n30 POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_n20 -PREHOOK: query: analyze table mv1_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_n20 -PREHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_n20 -POSTHOOK: Output: default@mv1_n20 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n30.empid, dependents_n20.empid, emps_n30.deptno from emps_n30 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt_2.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt_2.q.out index 61b0efc059..d9a48fb4b9 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt_2.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_no_join_opt_2.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n30.deptno SCRIPT [] POSTHOOK: Lineage: emps_n30.empid SCRIPT [] POSTHOOK: Lineage: emps_n30.name SCRIPT [] POSTHOOK: Lineage: emps_n30.salary SCRIPT [] -PREHOOK: query: analyze table emps_n30 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n30 -PREHOOK: Output: default@emps_n30 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n30 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n30 -POSTHOOK: Output: default@emps_n30 -#### A masked pattern was here #### PREHOOK: query: create table depts_n20 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n20 POSTHOOK: Lineage: depts_n20.deptno SCRIPT [] POSTHOOK: Lineage: depts_n20.locationid SCRIPT [] POSTHOOK: Lineage: depts_n20.name SCRIPT [] -PREHOOK: query: analyze table depts_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n20 -PREHOOK: Output: default@depts_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n20 -POSTHOOK: Output: default@depts_n20 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n20 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n20 POSTHOOK: Lineage: dependents_n20.empid SCRIPT [] POSTHOOK: Lineage: dependents_n20.name SCRIPT [] -PREHOOK: query: analyze table dependents_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n20 -PREHOOK: Output: default@dependents_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n20 -POSTHOOK: Output: default@dependents_n20 -#### A masked pattern was here #### PREHOOK: query: create table locations_n20 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n20 POSTHOOK: Lineage: locations_n20.locationid SCRIPT [] POSTHOOK: Lineage: locations_n20.name SCRIPT [] -PREHOOK: query: analyze table locations_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n20 -PREHOOK: Output: default@locations_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n20 -POSTHOOK: Output: default@locations_n20 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely @@ -190,20 +150,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).commission SIMPLE [(emps_n30 POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).empid SIMPLE [(emps_n30)emps_n30.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).name SIMPLE [(emps_n30)emps_n30.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@deptno=10 -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@deptno=10 -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### PREHOOK: query: explain select * from (select * from emps_n30 where empid < 120) t @@ -369,24 +315,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).salary SIMPLE [(emps_n30)emp POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=20).commission SIMPLE [(emps_n30)emps_n30.FieldSchema(name:commission, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=20).name SIMPLE [(emps_n30)emps_n30.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=20).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@deptno=10 -PREHOOK: Input: default@mv1_part_n2@deptno=20 -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@deptno=10 -PREHOOK: Output: default@mv1_part_n2@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@deptno=10 -POSTHOOK: Input: default@mv1_part_n2@deptno=20 -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@deptno=10 -POSTHOOK: Output: default@mv1_part_n2@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n30.name, emps_n30.salary, emps_n30.commission from emps_n30 @@ -546,20 +474,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).commission SIMPLE [(emps_n30 POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).empid SIMPLE [(emps_n30)emps_n30.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).name SIMPLE [(emps_n30)emps_n30.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@deptno=10 -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@deptno=10 -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### PREHOOK: query: explain select * from emps_n30 where empid > 120 union all select * from emps_n30 where empid < 150 @@ -690,32 +604,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Bill).salary SIMPLE [(emps_n30)emp POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Eric).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Eric -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Eric -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Eric -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Eric -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name, salary from emps_n30 group by name, salary PREHOOK: type: QUERY @@ -802,32 +690,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Bill).salary SIMPLE [(emps_n30)emp POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Eric).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Eric -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Eric -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Eric -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Eric -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n30 group by name PREHOOK: type: QUERY @@ -952,28 +814,6 @@ POSTHOOK: Output: default@mv1_part_n2@name=Theodore POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Bill).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n30 where deptno = 10 group by name PREHOOK: type: QUERY @@ -1105,32 +945,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n3 POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).c EXPRESSION [(emps_n30)emps_n30.null, ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).s EXPRESSION [(emps_n30)emps_n30.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Eric -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Eric -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Eric -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Eric -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n30 group by name PREHOOK: type: QUERY diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_1.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_1.q.out index 68f4947137..786e2c71a1 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_1.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_1.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n30.deptno SCRIPT [] POSTHOOK: Lineage: emps_n30.empid SCRIPT [] POSTHOOK: Lineage: emps_n30.name SCRIPT [] POSTHOOK: Lineage: emps_n30.salary SCRIPT [] -PREHOOK: query: analyze table emps_n30 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n30 -PREHOOK: Output: default@emps_n30 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n30 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n30 -POSTHOOK: Output: default@emps_n30 -#### A masked pattern was here #### PREHOOK: query: create table depts_n20 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n20 POSTHOOK: Lineage: depts_n20.deptno SCRIPT [] POSTHOOK: Lineage: depts_n20.locationid SCRIPT [] POSTHOOK: Lineage: depts_n20.name SCRIPT [] -PREHOOK: query: analyze table depts_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n20 -PREHOOK: Output: default@depts_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n20 -POSTHOOK: Output: default@depts_n20 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n20 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n20 POSTHOOK: Lineage: dependents_n20.empid SCRIPT [] POSTHOOK: Lineage: dependents_n20.name SCRIPT [] -PREHOOK: query: analyze table dependents_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n20 -PREHOOK: Output: default@dependents_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n20 -POSTHOOK: Output: default@dependents_n20 -#### A masked pattern was here #### PREHOOK: query: create table locations_n20 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n20 POSTHOOK: Lineage: locations_n20.locationid SCRIPT [] POSTHOOK: Lineage: locations_n20.name SCRIPT [] -PREHOOK: query: analyze table locations_n20 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n20 -PREHOOK: Output: default@locations_n20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n20 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n20 -POSTHOOK: Output: default@locations_n20 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n30 add constraint pk1 primary key (empid) disable novalidate rely @@ -190,20 +150,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).commission SIMPLE [(emps_n30 POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).empid SIMPLE [(emps_n30)emps_n30.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).name SIMPLE [(emps_n30)emps_n30.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@deptno=10 -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@deptno=10 -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### PREHOOK: query: explain select * from (select * from emps_n30 where empid < 120) t @@ -369,24 +315,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).salary SIMPLE [(emps_n30)emp POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=20).commission SIMPLE [(emps_n30)emps_n30.FieldSchema(name:commission, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=20).name SIMPLE [(emps_n30)emps_n30.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=20).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@deptno=10 -PREHOOK: Input: default@mv1_part_n2@deptno=20 -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@deptno=10 -PREHOOK: Output: default@mv1_part_n2@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@deptno=10 -POSTHOOK: Input: default@mv1_part_n2@deptno=20 -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@deptno=10 -POSTHOOK: Output: default@mv1_part_n2@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select emps_n30.name, emps_n30.salary, emps_n30.commission from emps_n30 @@ -476,20 +404,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).commission SIMPLE [(emps_n30 POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).empid SIMPLE [(emps_n30)emps_n30.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).name SIMPLE [(emps_n30)emps_n30.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(deptno=10).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@deptno=10 -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@deptno=10 -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@deptno=10 -#### A masked pattern was here #### PREHOOK: query: explain select * from emps_n30 where empid > 120 union all select * from emps_n30 where empid < 150 @@ -620,32 +534,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Bill).salary SIMPLE [(emps_n30)emp POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Eric).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Eric -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Eric -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Eric -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Eric -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name, salary from emps_n30 group by name, salary PREHOOK: type: QUERY @@ -732,32 +620,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Bill).salary SIMPLE [(emps_n30)emp POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Eric).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Eric -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Eric -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Eric -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Eric -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n30 group by name PREHOOK: type: QUERY @@ -882,28 +744,6 @@ POSTHOOK: Output: default@mv1_part_n2@name=Theodore POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Bill).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n30 where deptno = 10 group by name PREHOOK: type: QUERY @@ -1035,32 +875,6 @@ POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Sebastian).salary SIMPLE [(emps_n3 POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).c EXPRESSION [(emps_n30)emps_n30.null, ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).s EXPRESSION [(emps_n30)emps_n30.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n2 PARTITION(name=Theodore).salary SIMPLE [(emps_n30)emps_n30.FieldSchema(name:salary, type:float, comment:null), ] -PREHOOK: query: analyze table mv1_part_n2 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n2 -PREHOOK: Input: default@mv1_part_n2@name=Bill -PREHOOK: Input: default@mv1_part_n2@name=Eric -PREHOOK: Input: default@mv1_part_n2@name=Sebastian -PREHOOK: Input: default@mv1_part_n2@name=Theodore -PREHOOK: Output: default@mv1_part_n2 -PREHOOK: Output: default@mv1_part_n2@name=Bill -PREHOOK: Output: default@mv1_part_n2@name=Eric -PREHOOK: Output: default@mv1_part_n2@name=Sebastian -PREHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n2 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n2 -POSTHOOK: Input: default@mv1_part_n2@name=Bill -POSTHOOK: Input: default@mv1_part_n2@name=Eric -POSTHOOK: Input: default@mv1_part_n2@name=Sebastian -POSTHOOK: Input: default@mv1_part_n2@name=Theodore -POSTHOOK: Output: default@mv1_part_n2 -POSTHOOK: Output: default@mv1_part_n2@name=Bill -POSTHOOK: Output: default@mv1_part_n2@name=Eric -POSTHOOK: Output: default@mv1_part_n2@name=Sebastian -POSTHOOK: Output: default@mv1_part_n2@name=Theodore -#### A masked pattern was here #### PREHOOK: query: explain select name from emps_n30 group by name PREHOOK: type: QUERY diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_2.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_2.q.out index 67a11c0296..9765fdb268 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_2.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_part_2.q.out @@ -33,16 +33,6 @@ POSTHOOK: Lineage: emps_n00.deptno SCRIPT [] POSTHOOK: Lineage: emps_n00.empid SCRIPT [] POSTHOOK: Lineage: emps_n00.name SCRIPT [] POSTHOOK: Lineage: emps_n00.salary SCRIPT [] -PREHOOK: query: analyze table emps_n00 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@emps_n00 -PREHOOK: Output: default@emps_n00 -#### A masked pattern was here #### -POSTHOOK: query: analyze table emps_n00 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@emps_n00 -POSTHOOK: Output: default@emps_n00 -#### A masked pattern was here #### PREHOOK: query: create table depts_n00 ( deptno int, name varchar(256), @@ -70,16 +60,6 @@ POSTHOOK: Output: default@depts_n00 POSTHOOK: Lineage: depts_n00.deptno SCRIPT [] POSTHOOK: Lineage: depts_n00.locationid SCRIPT [] POSTHOOK: Lineage: depts_n00.name SCRIPT [] -PREHOOK: query: analyze table depts_n00 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@depts_n00 -PREHOOK: Output: default@depts_n00 -#### A masked pattern was here #### -POSTHOOK: query: analyze table depts_n00 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@depts_n00 -POSTHOOK: Output: default@depts_n00 -#### A masked pattern was here #### PREHOOK: query: create table dependents_n00 ( empid int, name varchar(256)) @@ -104,16 +84,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@dependents_n00 POSTHOOK: Lineage: dependents_n00.empid SCRIPT [] POSTHOOK: Lineage: dependents_n00.name SCRIPT [] -PREHOOK: query: analyze table dependents_n00 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dependents_n00 -PREHOOK: Output: default@dependents_n00 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dependents_n00 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dependents_n00 -POSTHOOK: Output: default@dependents_n00 -#### A masked pattern was here #### PREHOOK: query: create table locations_n00 ( locationid int, name varchar(256)) @@ -138,16 +108,6 @@ POSTHOOK: Input: _dummy_database@_dummy_table POSTHOOK: Output: default@locations_n00 POSTHOOK: Lineage: locations_n00.locationid SCRIPT [] POSTHOOK: Lineage: locations_n00.name SCRIPT [] -PREHOOK: query: analyze table locations_n00 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@locations_n00 -PREHOOK: Output: default@locations_n00 -#### A masked pattern was here #### -POSTHOOK: query: analyze table locations_n00 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@locations_n00 -POSTHOOK: Output: default@locations_n00 -#### A masked pattern was here #### PREHOOK: query: alter table emps_n00 add constraint pk1 primary key (empid) disable novalidate rely PREHOOK: type: ALTERTABLE_ADDCONSTRAINT POSTHOOK: query: alter table emps_n00 add constraint pk1 primary key (empid) disable novalidate rely @@ -193,20 +153,6 @@ POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_part_n0 POSTHOOK: Output: default@mv1_part_n0@deptno=20 POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=20).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] -PREHOOK: query: analyze table mv1_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n0 -PREHOOK: Input: default@mv1_part_n0@deptno=20 -PREHOOK: Output: default@mv1_part_n0 -PREHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n0 -POSTHOOK: Input: default@mv1_part_n0@deptno=20 -POSTHOOK: Output: default@mv1_part_n0 -POSTHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n00 join depts_n00 using (deptno) where depts_n00.deptno >= 20 @@ -291,20 +237,6 @@ POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_part_n0 POSTHOOK: Output: default@mv1_part_n0@deptno=20 POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=20).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] -PREHOOK: query: analyze table mv1_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n0 -PREHOOK: Input: default@mv1_part_n0@deptno=20 -PREHOOK: Output: default@mv1_part_n0 -PREHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n0 -POSTHOOK: Input: default@mv1_part_n0@deptno=20 -POSTHOOK: Output: default@mv1_part_n0 -POSTHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n00 join depts_n00 using (deptno) where depts_n00.deptno >= 20 @@ -389,20 +321,6 @@ POSTHOOK: Output: database:default POSTHOOK: Output: default@mv1_part_n0 POSTHOOK: Output: default@mv1_part_n0@deptno=20 POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=20).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] -PREHOOK: query: analyze table mv1_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n0 -PREHOOK: Input: default@mv1_part_n0@deptno=20 -PREHOOK: Output: default@mv1_part_n0 -PREHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n0 -POSTHOOK: Input: default@mv1_part_n0@deptno=20 -POSTHOOK: Output: default@mv1_part_n0 -POSTHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select empid from emps_n00 join depts_n00 using (deptno) where depts_n00.deptno >= 20 @@ -489,24 +407,6 @@ POSTHOOK: Output: default@mv1_part_n0@deptno=10 POSTHOOK: Output: default@mv1_part_n0@deptno=20 POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=10).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=20).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] -PREHOOK: query: analyze table mv1_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n0 -PREHOOK: Input: default@mv1_part_n0@deptno=10 -PREHOOK: Input: default@mv1_part_n0@deptno=20 -PREHOOK: Output: default@mv1_part_n0 -PREHOOK: Output: default@mv1_part_n0@deptno=10 -PREHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n0 -POSTHOOK: Input: default@mv1_part_n0@deptno=10 -POSTHOOK: Input: default@mv1_part_n0@deptno=20 -POSTHOOK: Output: default@mv1_part_n0 -POSTHOOK: Output: default@mv1_part_n0@deptno=10 -POSTHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n00.deptno from depts_n00 join emps_n00 using (deptno) where emps_n00.empid > 15 @@ -603,24 +503,6 @@ POSTHOOK: Output: default@mv1_part_n0@deptno=10 POSTHOOK: Output: default@mv1_part_n0@deptno=20 POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=10).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno=20).empid SIMPLE [(emps_n00)emps_n00.FieldSchema(name:empid, type:int, comment:null), ] -PREHOOK: query: analyze table mv1_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n0 -PREHOOK: Input: default@mv1_part_n0@deptno=10 -PREHOOK: Input: default@mv1_part_n0@deptno=20 -PREHOOK: Output: default@mv1_part_n0 -PREHOOK: Output: default@mv1_part_n0@deptno=10 -PREHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n0 -POSTHOOK: Input: default@mv1_part_n0@deptno=10 -POSTHOOK: Input: default@mv1_part_n0@deptno=20 -POSTHOOK: Output: default@mv1_part_n0 -POSTHOOK: Output: default@mv1_part_n0@deptno=10 -POSTHOOK: Output: default@mv1_part_n0@deptno=20 -#### A masked pattern was here #### PREHOOK: query: explain select depts_n00.deptno from depts_n00 join emps_n00 using (deptno) where emps_n00.empid > 15 @@ -780,7 +662,7 @@ POSTHOOK: query: drop materialized view mv1_part_n0 POSTHOOK: type: DROP_MATERIALIZED_VIEW POSTHOOK: Input: default@mv1_part_n0 POSTHOOK: Output: default@mv1_part_n0 -Warning: Shuffle Join MERGEJOIN[19][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[29][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 2' is a cross product PREHOOK: query: create materialized view mv1_part_n0 partitioned on (deptno2) as select depts_n00.name, dependents_n00.name as name2, emps_n00.deptno, depts_n00.deptno as deptno2, dependents_n00.empid from depts_n00, dependents_n00, emps_n00 @@ -814,24 +696,6 @@ POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno2=30).deptno SIMPLE [(emps_n00)em POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno2=30).empid SIMPLE [(dependents_n00)dependents_n00.FieldSchema(name:empid, type:int, comment:null), ] POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno2=30).name SIMPLE [(depts_n00)depts_n00.FieldSchema(name:name, type:varchar(256), comment:null), ] POSTHOOK: Lineage: mv1_part_n0 PARTITION(deptno2=30).name2 SIMPLE [(dependents_n00)dependents_n00.FieldSchema(name:name, type:varchar(256), comment:null), ] -PREHOOK: query: analyze table mv1_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@mv1_part_n0 -PREHOOK: Input: default@mv1_part_n0@deptno2=20 -PREHOOK: Input: default@mv1_part_n0@deptno2=30 -PREHOOK: Output: default@mv1_part_n0 -PREHOOK: Output: default@mv1_part_n0@deptno2=20 -PREHOOK: Output: default@mv1_part_n0@deptno2=30 -#### A masked pattern was here #### -POSTHOOK: query: analyze table mv1_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@mv1_part_n0 -POSTHOOK: Input: default@mv1_part_n0@deptno2=20 -POSTHOOK: Input: default@mv1_part_n0@deptno2=30 -POSTHOOK: Output: default@mv1_part_n0 -POSTHOOK: Output: default@mv1_part_n0@deptno2=20 -POSTHOOK: Output: default@mv1_part_n0@deptno2=30 -#### A masked pattern was here #### PREHOOK: query: explain select dependents_n00.empid from depts_n00 diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb.q.out index f6aadd5b13..d98a9b665d 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb.q.out @@ -528,56 +528,6 @@ POSTHOOK: Lineage: lineorder_n0.lo_shippriority SIMPLE [(lineorder_ext_n0)lineor POSTHOOK: Lineage: lineorder_n0.lo_suppkey SIMPLE [(lineorder_ext_n0)lineorder_ext_n0.FieldSchema(name:lo_suppkey, type:bigint, comment:null), ] POSTHOOK: Lineage: lineorder_n0.lo_supplycost SIMPLE [(lineorder_ext_n0)lineorder_ext_n0.FieldSchema(name:lo_supplycost, type:double, comment:null), ] POSTHOOK: Lineage: lineorder_n0.lo_tax SIMPLE [(lineorder_ext_n0)lineorder_ext_n0.FieldSchema(name:lo_tax, type:double, comment:null), ] -PREHOOK: query: analyze table customer_n1 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@customer_n1 -PREHOOK: Output: default@customer_n1 -#### A masked pattern was here #### -POSTHOOK: query: analyze table customer_n1 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@customer_n1 -POSTHOOK: Output: default@customer_n1 -#### A masked pattern was here #### -PREHOOK: query: analyze table dates_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dates_n0 -PREHOOK: Output: default@dates_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table dates_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dates_n0 -POSTHOOK: Output: default@dates_n0 -#### A masked pattern was here #### -PREHOOK: query: analyze table ssb_part_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@ssb_part_n0 -PREHOOK: Output: default@ssb_part_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table ssb_part_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@ssb_part_n0 -POSTHOOK: Output: default@ssb_part_n0 -#### A masked pattern was here #### -PREHOOK: query: analyze table supplier_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@supplier_n0 -PREHOOK: Output: default@supplier_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table supplier_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@supplier_n0 -POSTHOOK: Output: default@supplier_n0 -#### A masked pattern was here #### -PREHOOK: query: analyze table lineorder_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@lineorder_n0 -PREHOOK: Output: default@lineorder_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table lineorder_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@lineorder_n0 -POSTHOOK: Output: default@lineorder_n0 -#### A masked pattern was here #### PREHOOK: query: CREATE MATERIALIZED VIEW `ssb_mv_n0` AS SELECT @@ -697,22 +647,22 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_year = 1993) and (lo_quantity < 25.0D) and lo_discount BETWEEN 1.0D AND 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((d_year = 1993) and (lo_quantity < 25.0D) and lo_discount BETWEEN 1.0D AND 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: discounted_price (type: double) outputColumnNames: discounted_price - Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(discounted_price) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -723,10 +673,10 @@ STAGE PLANS: aggregations: sum(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -785,22 +735,22 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_yearmonthnum = 199401) and lo_discount BETWEEN 4.0D AND 6.0D and lo_quantity BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((d_yearmonthnum = 199401) and lo_discount BETWEEN 4.0D AND 6.0D and lo_quantity BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: discounted_price (type: double) outputColumnNames: discounted_price - Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(discounted_price) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -811,10 +761,10 @@ STAGE PLANS: aggregations: sum(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -875,22 +825,22 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_year = 1994) and (d_weeknuminyear = 6) and lo_discount BETWEEN 5.0D AND 7.0D and lo_quantity BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((d_weeknuminyear = 6) and (d_year = 1994) and lo_discount BETWEEN 5.0D AND 7.0D and lo_quantity BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: discounted_price (type: double) outputColumnNames: discounted_price - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(discounted_price) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 40 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 40 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -901,10 +851,10 @@ STAGE PLANS: aggregations: sum(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 40 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 40 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -978,25 +928,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((p_category = 'MFGR#12') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((p_category = 'MFGR#12') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: d_year (type: int), p_brand1 (type: string), lo_revenue (type: double) outputColumnNames: d_year, p_brand1, lo_revenue - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int), p_brand1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1008,15 +958,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col2 (type: double), _col0 (type: int), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: int), _col2 (type: string) sort order: ++ - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1024,10 +974,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: double), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1101,25 +1051,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((s_region = 'ASIA') and p_brand1 BETWEEN 'MFGR#2221' AND 'MFGR#2228') (type: boolean) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((s_region = 'ASIA') and p_brand1 BETWEEN 'MFGR#2221' AND 'MFGR#2228') (type: boolean) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: d_year (type: int), p_brand1 (type: string), lo_revenue (type: double) outputColumnNames: d_year, p_brand1, lo_revenue - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int), p_brand1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1131,15 +1081,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col2 (type: double), _col0 (type: int), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: int), _col2 (type: string) sort order: ++ - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1147,10 +1097,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: double), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1224,25 +1174,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((p_brand1 = 'MFGR#2239') and (s_region = 'EUROPE')) (type: boolean) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((p_brand1 = 'MFGR#2239') and (s_region = 'EUROPE')) (type: boolean) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: d_year (type: int), lo_revenue (type: double) outputColumnNames: d_year, lo_revenue - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1254,15 +1204,15 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: double), _col0 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: int) sort order: + - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1270,10 +1220,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: double), KEY.reducesinkkey0 (type: int), 'MFGR#2239' (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 105 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 105 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1351,25 +1301,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_year >= 1992) and (c_region = 'ASIA') and (s_region = 'ASIA') and (d_year <= 1997)) (type: boolean) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_region = 'ASIA') and (d_year <= 1997) and (d_year >= 1992) and (s_region = 'ASIA')) (type: boolean) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_nation (type: string), d_year (type: int), s_nation (type: string), lo_revenue (type: double) outputColumnNames: c_nation, d_year, s_nation, lo_revenue - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int), c_nation (type: string), s_nation (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1381,15 +1331,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1397,10 +1347,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1476,25 +1426,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_year >= 1992) and (c_nation = 'UNITED STATES') and (s_nation = 'UNITED STATES') and (d_year <= 1997)) (type: boolean) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_nation = 'UNITED STATES') and (d_year <= 1997) and (d_year >= 1992) and (s_nation = 'UNITED STATES')) (type: boolean) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_city (type: string), d_year (type: int), s_city (type: string), lo_revenue (type: double) outputColumnNames: c_city, d_year, s_city, lo_revenue - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int), c_city (type: string), s_city (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1506,15 +1456,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1522,10 +1472,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1601,21 +1551,21 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (s_city) IN ('UNITED KI1', 'UNITED KI5') and (d_year >= 1992) and (d_year <= 1997)) (type: boolean) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (d_year <= 1997) and (d_year >= 1992) and (s_city) IN ('UNITED KI1', 'UNITED KI5')) (type: boolean) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int), c_city (type: string), s_city (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1627,15 +1577,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1643,10 +1593,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1722,25 +1672,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (s_city) IN ('UNITED KI1', 'UNITED KI5') and (d_yearmonth = 'Dec1997')) (type: boolean) - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (d_yearmonth = 'Dec1997') and (s_city) IN ('UNITED KI1', 'UNITED KI5')) (type: boolean) - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_city (type: string), d_year (type: int), s_city (type: string), lo_revenue (type: double) outputColumnNames: c_city, d_year, s_city, lo_revenue - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(lo_revenue) keys: d_year (type: int), c_city (type: string), s_city (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1752,15 +1702,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1768,10 +1718,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 564 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1853,25 +1803,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((p_mfgr) IN ('MFGR#1', 'MFGR#2') and (c_region = 'AMERICA') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_region = 'AMERICA') and (p_mfgr) IN ('MFGR#1', 'MFGR#2') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_nation (type: string), d_year (type: int), net_revenue (type: double) outputColumnNames: c_nation, d_year, net_revenue - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(net_revenue) keys: d_year (type: int), c_nation (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1883,11 +1833,11 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1895,10 +1845,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 748 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1982,25 +1932,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_year) IN (1997, 1998) and (p_mfgr) IN ('MFGR#1', 'MFGR#2') and (c_region = 'AMERICA') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_region = 'AMERICA') and (d_year) IN (1997, 1998) and (p_mfgr) IN ('MFGR#1', 'MFGR#2') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: d_year (type: int), p_category (type: string), s_nation (type: string), net_revenue (type: double) outputColumnNames: d_year, p_category, s_nation, net_revenue - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(net_revenue) keys: d_year (type: int), s_nation (type: string), p_category (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -2012,11 +1962,11 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -2024,10 +1974,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), VALUE._col0 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2111,25 +2061,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv_n0 filterExpr: ((d_year) IN (1997, 1998) and (c_region = 'AMERICA') and (p_category = 'MFGR#14') and (s_nation = 'UNITED STATES')) (type: boolean) - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_region = 'AMERICA') and (d_year) IN (1997, 1998) and (p_category = 'MFGR#14') and (s_nation = 'UNITED STATES')) (type: boolean) - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: d_year (type: int), p_brand1 (type: string), s_city (type: string), net_revenue (type: double) outputColumnNames: d_year, p_brand1, s_city, net_revenue - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(net_revenue) keys: d_year (type: int), s_city (type: string), p_brand1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -2141,11 +2091,11 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -2153,10 +2103,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), VALUE._col0 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 932 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat diff --git a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb_2.q.out b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb_2.q.out index 6f7dcdb66b..abefa7eaa2 100644 --- a/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb_2.q.out +++ b/ql/src/test/results/clientpositive/llap/materialized_view_rewrite_ssb_2.q.out @@ -528,56 +528,6 @@ POSTHOOK: Lineage: lineorder.lo_shippriority SIMPLE [(lineorder_ext)lineorder_ex POSTHOOK: Lineage: lineorder.lo_suppkey SIMPLE [(lineorder_ext)lineorder_ext.FieldSchema(name:lo_suppkey, type:bigint, comment:null), ] POSTHOOK: Lineage: lineorder.lo_supplycost SIMPLE [(lineorder_ext)lineorder_ext.FieldSchema(name:lo_supplycost, type:double, comment:null), ] POSTHOOK: Lineage: lineorder.lo_tax SIMPLE [(lineorder_ext)lineorder_ext.FieldSchema(name:lo_tax, type:double, comment:null), ] -PREHOOK: query: analyze table customer_n0 compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@customer_n0 -PREHOOK: Output: default@customer_n0 -#### A masked pattern was here #### -POSTHOOK: query: analyze table customer_n0 compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@customer_n0 -POSTHOOK: Output: default@customer_n0 -#### A masked pattern was here #### -PREHOOK: query: analyze table dates compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@dates -PREHOOK: Output: default@dates -#### A masked pattern was here #### -POSTHOOK: query: analyze table dates compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@dates -POSTHOOK: Output: default@dates -#### A masked pattern was here #### -PREHOOK: query: analyze table ssb_part compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@ssb_part -PREHOOK: Output: default@ssb_part -#### A masked pattern was here #### -POSTHOOK: query: analyze table ssb_part compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@ssb_part -POSTHOOK: Output: default@ssb_part -#### A masked pattern was here #### -PREHOOK: query: analyze table supplier compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@supplier -PREHOOK: Output: default@supplier -#### A masked pattern was here #### -POSTHOOK: query: analyze table supplier compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@supplier -POSTHOOK: Output: default@supplier -#### A masked pattern was here #### -PREHOOK: query: analyze table lineorder compute statistics for columns -PREHOOK: type: ANALYZE_TABLE -PREHOOK: Input: default@lineorder -PREHOOK: Output: default@lineorder -#### A masked pattern was here #### -POSTHOOK: query: analyze table lineorder compute statistics for columns -POSTHOOK: type: ANALYZE_TABLE -POSTHOOK: Input: default@lineorder -POSTHOOK: Output: default@lineorder -#### A masked pattern was here #### PREHOOK: query: CREATE MATERIALIZED VIEW `ssb_mv` AS SELECT @@ -699,22 +649,22 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_year) = 1993) and (UDFToDouble(lo_quantity) < 25.0D) and UDFToDouble(lo_discount) BETWEEN 1.0D AND 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToDouble(lo_quantity) < 25.0D) and (UDFToInteger(d_year) = 1993) and UDFToDouble(lo_discount) BETWEEN 1.0D AND 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: discounted_price (type: double) outputColumnNames: discounted_price - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(discounted_price) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -725,10 +675,10 @@ STAGE PLANS: aggregations: sum(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -787,22 +737,22 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_yearmonthnum) = 199401) and UDFToDouble(lo_discount) BETWEEN 4.0D AND 6.0D and UDFToDouble(lo_quantity) BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_yearmonthnum) = 199401) and UDFToDouble(lo_discount) BETWEEN 4.0D AND 6.0D and UDFToDouble(lo_quantity) BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: discounted_price (type: double) outputColumnNames: discounted_price - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(discounted_price) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -813,10 +763,10 @@ STAGE PLANS: aggregations: sum(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 568 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -877,22 +827,22 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_year) = 1994) and (UDFToInteger(d_weeknuminyear) = 6) and UDFToDouble(lo_discount) BETWEEN 5.0D AND 7.0D and UDFToDouble(lo_quantity) BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_weeknuminyear) = 6) and (UDFToInteger(d_year) = 1994) and UDFToDouble(lo_discount) BETWEEN 5.0D AND 7.0D and UDFToDouble(lo_quantity) BETWEEN 26.0D AND 35.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: discounted_price (type: double) outputColumnNames: discounted_price - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(discounted_price) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 752 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 752 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -903,10 +853,10 @@ STAGE PLANS: aggregations: sum(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 752 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 752 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -980,25 +930,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((p_category = 'MFGR#12') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((p_category = 'MFGR#12') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: UDFToInteger(d_year) (type: int), p_brand1 (type: string), lo_revenue (type: double) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col2) keys: _col0 (type: int), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1010,15 +960,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col2 (type: double), _col0 (type: int), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: int), _col2 (type: string) sort order: ++ - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1026,10 +976,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: double), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1103,25 +1053,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((s_region = 'ASIA') and p_brand1 BETWEEN 'MFGR#2221' AND 'MFGR#2228') (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((s_region = 'ASIA') and p_brand1 BETWEEN 'MFGR#2221' AND 'MFGR#2228') (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: UDFToInteger(d_year) (type: int), p_brand1 (type: string), lo_revenue (type: double) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col2) keys: _col0 (type: int), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1133,15 +1083,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col2 (type: double), _col0 (type: int), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: int), _col2 (type: string) sort order: ++ - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1149,10 +1099,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: double), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1226,25 +1176,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((p_brand1 = 'MFGR#2239') and (s_region = 'EUROPE')) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((p_brand1 = 'MFGR#2239') and (s_region = 'EUROPE')) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: UDFToInteger(d_year) (type: int), lo_revenue (type: double) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col1) keys: _col0 (type: int) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1256,15 +1206,15 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: double), _col0 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col1 (type: int) sort order: + - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1272,10 +1222,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: double), KEY.reducesinkkey0 (type: int), 'MFGR#2239' (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 105 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 105 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1353,25 +1303,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_year) >= 1992) and (c_region = 'ASIA') and (s_region = 'ASIA') and (UDFToInteger(d_year) <= 1997)) (type: boolean) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_year) <= 1997) and (UDFToInteger(d_year) >= 1992) and (c_region = 'ASIA') and (s_region = 'ASIA')) (type: boolean) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_nation (type: string), s_nation (type: string), UDFToInteger(d_year) (type: int), lo_revenue (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col3) keys: _col2 (type: int), _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1383,15 +1333,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1399,10 +1349,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1478,25 +1428,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_year) >= 1992) and (c_nation = 'UNITED STATES') and (s_nation = 'UNITED STATES') and (UDFToInteger(d_year) <= 1997)) (type: boolean) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_year) <= 1997) and (UDFToInteger(d_year) >= 1992) and (c_nation = 'UNITED STATES') and (s_nation = 'UNITED STATES')) (type: boolean) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_city (type: string), s_city (type: string), UDFToInteger(d_year) (type: int), lo_revenue (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col3) keys: _col2 (type: int), _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1508,15 +1458,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1524,10 +1474,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1603,25 +1553,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (s_city) IN ('UNITED KI1', 'UNITED KI5') and (UDFToInteger(d_year) >= 1992) and (UDFToInteger(d_year) <= 1997)) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_year) <= 1997) and (UDFToInteger(d_year) >= 1992) and (c_city) IN ('UNITED KI1', 'UNITED KI5') and (s_city) IN ('UNITED KI1', 'UNITED KI5')) (type: boolean) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_city (type: string), s_city (type: string), UDFToInteger(d_year) (type: int), lo_revenue (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col3) keys: _col2 (type: int), _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1633,15 +1583,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1649,10 +1599,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1728,25 +1678,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (s_city) IN ('UNITED KI1', 'UNITED KI5') and (d_yearmonth = 'Dec1997')) (type: boolean) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_city) IN ('UNITED KI1', 'UNITED KI5') and (d_yearmonth = 'Dec1997') and (s_city) IN ('UNITED KI1', 'UNITED KI5')) (type: boolean) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: c_city (type: string), s_city (type: string), UDFToInteger(d_year) (type: int), lo_revenue (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 344 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col3) keys: _col2 (type: int), _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1758,15 +1708,15 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col1 (type: string), _col2 (type: string), _col0 (type: int), _col3 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col2 (type: int), _col3 (type: double) sort order: +- - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col0 (type: string), _col1 (type: string) Reducer 3 Execution mode: vectorized, llap @@ -1774,10 +1724,10 @@ STAGE PLANS: Select Operator expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1859,25 +1809,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((p_mfgr) IN ('MFGR#1', 'MFGR#2') and (c_region = 'AMERICA') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((c_region = 'AMERICA') and (p_mfgr) IN ('MFGR#1', 'MFGR#2') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: UDFToInteger(d_year) (type: int), c_nation (type: string), net_revenue (type: double) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 428 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col2) keys: _col0 (type: int), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -1889,11 +1839,11 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -1901,10 +1851,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 928 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1988,25 +1938,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_year)) IN (1997, 1998) and (p_mfgr) IN ('MFGR#1', 'MFGR#2') and (c_region = 'AMERICA') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 512 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_year)) IN (1997, 1998) and (c_region = 'AMERICA') and (p_mfgr) IN ('MFGR#1', 'MFGR#2') and (s_region = 'AMERICA')) (type: boolean) - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 512 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: UDFToInteger(d_year) (type: int), s_nation (type: string), p_category (type: string), net_revenue (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 512 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col3) keys: _col0 (type: int), _col1 (type: string), _col2 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -2018,11 +1968,11 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -2030,10 +1980,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), VALUE._col0 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2117,25 +2067,25 @@ STAGE PLANS: TableScan alias: default.ssb_mv filterExpr: ((UDFToInteger(d_year)) IN (1997, 1998) and (c_region = 'AMERICA') and (p_category = 'MFGR#14') and (s_nation = 'UNITED STATES')) (type: boolean) - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 512 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((UDFToInteger(d_year)) IN (1997, 1998) and (c_region = 'AMERICA') and (p_category = 'MFGR#14') and (s_nation = 'UNITED STATES')) (type: boolean) - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 512 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: UDFToInteger(d_year) (type: int), s_city (type: string), p_brand1 (type: string), net_revenue (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 512 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: sum(_col3) keys: _col0 (type: int), _col1 (type: string), _col2 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Execution mode: vectorized, llap LLAP IO: all inputs @@ -2147,11 +2097,11 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: string), KEY._col2 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) sort order: +++ - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col3 (type: double) Reducer 3 Execution mode: vectorized, llap @@ -2159,10 +2109,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), VALUE._col0 (type: double) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat