From 9887ff6cafffba5cf1b7648f4a3ba13f004320ed Mon Sep 17 00:00:00 2001 From: Ashutosh Chauhan Date: Fri, 22 May 2015 18:26:52 -0700 Subject: [PATCH] HIVE-10807 : Invalidate basic stats for insert queries if autogather=false --- .../org/apache/hadoop/hive/ql/QueryProperties.java | 10 -- .../org/apache/hadoop/hive/ql/exec/StatsTask.java | 8 +- .../hadoop/hive/ql/optimizer/GenMRFileSink1.java | 1 + .../hadoop/hive/ql/optimizer/GenMapRedUtils.java | 66 ++++----- .../apache/hadoop/hive/ql/parse/QBParseInfo.java | 9 -- .../hadoop/hive/ql/parse/SemanticAnalyzer.java | 10 +- ql/src/test/queries/clientpositive/insert_into1.q | 12 +- .../stats_partialscan_autogether.q.out | 2 +- .../results/clientpositive/auto_join_filters.q.out | 4 +- .../results/clientpositive/auto_join_nulls.q.out | 6 +- .../clientpositive/auto_sortmerge_join_1.q.out | 60 ++++---- .../clientpositive/auto_sortmerge_join_11.q.out | 64 ++++----- .../clientpositive/auto_sortmerge_join_12.q.out | 32 ++--- .../clientpositive/auto_sortmerge_join_2.q.out | 52 +++---- .../clientpositive/auto_sortmerge_join_3.q.out | 48 +++---- .../clientpositive/auto_sortmerge_join_4.q.out | 48 +++---- .../clientpositive/auto_sortmerge_join_5.q.out | 48 +++++-- .../clientpositive/auto_sortmerge_join_7.q.out | 72 +++++----- .../clientpositive/auto_sortmerge_join_8.q.out | 72 +++++----- .../results/clientpositive/bucket_map_join_1.q.out | 12 +- .../results/clientpositive/bucket_map_join_2.q.out | 12 +- .../clientpositive/bucket_map_join_spark1.q.out | 24 ++-- .../clientpositive/bucket_map_join_spark2.q.out | 24 ++-- .../clientpositive/bucket_map_join_spark3.q.out | 24 ++-- .../results/clientpositive/bucketcontext_1.q.out | 20 +-- .../results/clientpositive/bucketcontext_2.q.out | 20 +-- .../results/clientpositive/bucketcontext_3.q.out | 16 +-- .../results/clientpositive/bucketcontext_4.q.out | 16 +-- .../results/clientpositive/bucketcontext_5.q.out | 16 ++- .../results/clientpositive/bucketcontext_6.q.out | 16 +-- .../results/clientpositive/bucketcontext_7.q.out | 24 ++-- .../results/clientpositive/bucketcontext_8.q.out | 24 ++-- .../results/clientpositive/bucketmapjoin1.q.out | 16 ++- .../results/clientpositive/bucketmapjoin10.q.out | 16 +-- .../results/clientpositive/bucketmapjoin11.q.out | 32 ++--- .../results/clientpositive/bucketmapjoin12.q.out | 16 +-- .../results/clientpositive/bucketmapjoin2.q.out | 28 ++-- .../results/clientpositive/bucketmapjoin3.q.out | 16 +-- .../results/clientpositive/bucketmapjoin4.q.out | 16 ++- .../results/clientpositive/bucketmapjoin5.q.out | 16 +-- .../results/clientpositive/bucketmapjoin7.q.out | 8 +- .../results/clientpositive/bucketmapjoin8.q.out | 16 +-- .../results/clientpositive/bucketmapjoin9.q.out | 16 +-- .../clientpositive/bucketmapjoin_negative.q.out | 12 +- .../clientpositive/bucketmapjoin_negative2.q.out | 16 ++- .../clientpositive/bucketmapjoin_negative3.q.out | 72 +++++++--- .../clientpositive/columnstats_partlvl.q.out | 8 +- .../clientpositive/columnstats_tbllvl.q.out | 16 ++- .../clientpositive/display_colstats_tbllvl.q.out | 8 +- .../encryption_insert_partition_dynamic.q.out | 12 +- .../encryption_insert_partition_static.q.out | 12 +- .../encryption_join_unencrypted_tbl.q.out | 12 +- .../results/clientpositive/groupby_sort_6.q.out | 4 +- .../test/results/clientpositive/insert_into1.q.out | 155 +++++++++++++++++++++ .../test/results/clientpositive/join_filters.q.out | 4 +- .../test/results/clientpositive/join_nulls.q.out | 6 +- .../clientpositive/list_bucket_dml_8.q.java1.7.out | 8 +- .../results/clientpositive/parquet_serde.q.out | 8 +- .../clientpositive/ql_rewrite_gbtoidx_cbo_2.q.out | 16 +-- .../clientpositive/skewjoin_union_remove_1.q.out | 20 ++- .../spark/auto_sortmerge_join_1.q.out | 28 ++-- .../spark/auto_sortmerge_join_12.q.out | 20 +-- .../spark/auto_sortmerge_join_2.q.out | 20 +-- .../spark/auto_sortmerge_join_3.q.out | 20 +-- .../spark/auto_sortmerge_join_4.q.out | 20 +-- .../spark/auto_sortmerge_join_5.q.out | 32 +++-- .../spark/auto_sortmerge_join_7.q.out | 32 ++--- .../spark/auto_sortmerge_join_8.q.out | 32 ++--- .../clientpositive/spark/bucket_map_join_1.q.out | 24 ++-- .../clientpositive/spark/bucket_map_join_2.q.out | 24 ++-- .../spark/bucket_map_join_spark1.q.out | 16 +-- .../spark/bucket_map_join_spark2.q.out | 16 +-- .../spark/bucket_map_join_spark3.q.out | 16 +-- .../clientpositive/spark/bucketmapjoin1.q.out | 24 ++-- .../clientpositive/spark/bucketmapjoin10.q.out | 16 +-- .../clientpositive/spark/bucketmapjoin11.q.out | 32 ++--- .../clientpositive/spark/bucketmapjoin12.q.out | 16 +-- .../clientpositive/spark/bucketmapjoin2.q.out | 28 ++-- .../clientpositive/spark/bucketmapjoin3.q.out | 16 +-- .../clientpositive/spark/bucketmapjoin4.q.out | 32 +++-- .../clientpositive/spark/bucketmapjoin5.q.out | 32 +++-- .../clientpositive/spark/bucketmapjoin7.q.out | 8 +- .../clientpositive/spark/bucketmapjoin8.q.out | 16 +-- .../clientpositive/spark/bucketmapjoin9.q.out | 16 +-- .../spark/bucketmapjoin_negative.q.out | 12 +- .../spark/bucketmapjoin_negative2.q.out | 16 ++- .../spark/bucketmapjoin_negative3.q.out | 144 ++++++++++++++----- .../test/results/clientpositive/spark/count.q.out | 10 +- .../clientpositive/spark/insert_into1.q.out | 120 ++++++++++++++++ .../spark/skewjoin_union_remove_1.q.out | 8 ++ .../test/results/clientpositive/spark/stats1.q.out | 6 +- .../results/clientpositive/spark/stats12.q.out | 8 +- .../results/clientpositive/spark/stats13.q.out | 8 +- .../results/clientpositive/spark/stats18.q.out | 4 +- .../test/results/clientpositive/spark/stats2.q.out | 4 + .../test/results/clientpositive/spark/stats3.q.out | 6 +- .../test/results/clientpositive/spark/stats6.q.out | 4 +- .../clientpositive/spark/stats_noscan_1.q.out | 8 +- .../clientpositive/spark/stats_noscan_2.q.out | 2 +- .../clientpositive/spark/stats_partscan_1_23.q.out | 6 +- .../clientpositive/spark/union_remove_1.q.out | 4 + .../clientpositive/spark/union_remove_10.q.out | 22 +-- .../clientpositive/spark/union_remove_11.q.out | 22 +-- .../clientpositive/spark/union_remove_12.q.out | 28 ++-- .../clientpositive/spark/union_remove_13.q.out | 28 ++-- .../clientpositive/spark/union_remove_14.q.out | 28 ++-- .../clientpositive/spark/union_remove_15.q.out | 4 + .../clientpositive/spark/union_remove_16.q.out | 22 +-- .../clientpositive/spark/union_remove_17.q.out | 4 + .../clientpositive/spark/union_remove_18.q.out | 4 + .../clientpositive/spark/union_remove_19.q.out | 12 ++ .../clientpositive/spark/union_remove_2.q.out | 4 + .../clientpositive/spark/union_remove_20.q.out | 4 + .../clientpositive/spark/union_remove_21.q.out | 4 + .../clientpositive/spark/union_remove_22.q.out | 8 ++ .../clientpositive/spark/union_remove_23.q.out | 4 + .../clientpositive/spark/union_remove_24.q.out | 4 + .../clientpositive/spark/union_remove_25.q.out | 18 ++- .../clientpositive/spark/union_remove_3.q.out | 4 + .../clientpositive/spark/union_remove_4.q.out | 22 +-- .../clientpositive/spark/union_remove_5.q.out | 22 +-- .../clientpositive/spark/union_remove_6.q.out | 8 ++ .../clientpositive/spark/union_remove_6_subq.q.out | 8 ++ .../clientpositive/spark/union_remove_7.q.out | 4 + .../clientpositive/spark/union_remove_8.q.out | 4 + .../clientpositive/spark/union_remove_9.q.out | 22 +-- ql/src/test/results/clientpositive/stats1.q.out | 6 +- ql/src/test/results/clientpositive/stats11.q.out | 32 +++-- ql/src/test/results/clientpositive/stats12.q.out | 8 +- ql/src/test/results/clientpositive/stats13.q.out | 8 +- ql/src/test/results/clientpositive/stats18.q.out | 4 +- ql/src/test/results/clientpositive/stats2.q.out | 4 + ql/src/test/results/clientpositive/stats3.q.out | 6 +- ql/src/test/results/clientpositive/stats6.q.out | 4 +- .../results/clientpositive/stats_noscan_1.q.out | 8 +- .../results/clientpositive/stats_noscan_2.q.out | 2 +- .../clientpositive/stats_partscan_1_23.q.out | 4 +- .../temp_table_display_colstats_tbllvl.q.out | 8 +- .../clientpositive/tez/auto_join_filters.q.out | 4 +- .../clientpositive/tez/auto_sortmerge_join_1.q.out | 36 ++--- .../tez/auto_sortmerge_join_11.q.out | 56 ++++---- .../tez/auto_sortmerge_join_12.q.out | 20 +-- .../clientpositive/tez/auto_sortmerge_join_2.q.out | 24 ++-- .../clientpositive/tez/auto_sortmerge_join_3.q.out | 36 ++--- .../clientpositive/tez/auto_sortmerge_join_4.q.out | 36 ++--- .../clientpositive/tez/auto_sortmerge_join_5.q.out | 48 +++++-- .../clientpositive/tez/auto_sortmerge_join_7.q.out | 48 +++---- .../clientpositive/tez/auto_sortmerge_join_8.q.out | 48 +++---- .../results/clientpositive/tez/explainuser_1.q.out | 6 +- .../results/clientpositive/tez/insert_into1.q.out | 124 +++++++++++++++++ .../clientpositive/tez/stats_noscan_1.q.out | 8 +- .../results/clientpositive/truncate_column.q.out | 34 ++--- .../truncate_column_list_bucket.q.out | 8 +- .../results/clientpositive/union_remove_1.q.out | 8 +- .../results/clientpositive/union_remove_10.q.out | 28 ++-- .../results/clientpositive/union_remove_11.q.out | 22 +-- .../results/clientpositive/union_remove_12.q.out | 30 ++-- .../results/clientpositive/union_remove_13.q.out | 30 ++-- .../results/clientpositive/union_remove_14.q.out | 30 ++-- .../results/clientpositive/union_remove_15.q.out | 8 +- .../results/clientpositive/union_remove_16.q.out | 26 ++-- .../results/clientpositive/union_remove_17.q.out | 4 + .../results/clientpositive/union_remove_18.q.out | 8 +- .../results/clientpositive/union_remove_19.q.out | 24 +++- .../results/clientpositive/union_remove_2.q.out | 10 +- .../results/clientpositive/union_remove_20.q.out | 8 +- .../results/clientpositive/union_remove_21.q.out | 8 +- .../results/clientpositive/union_remove_22.q.out | 16 ++- .../results/clientpositive/union_remove_23.q.out | 10 +- .../results/clientpositive/union_remove_24.q.out | 8 +- .../results/clientpositive/union_remove_25.q.out | 30 ++-- .../results/clientpositive/union_remove_3.q.out | 4 + .../results/clientpositive/union_remove_4.q.out | 26 ++-- .../results/clientpositive/union_remove_5.q.out | 28 ++-- .../results/clientpositive/union_remove_6.q.out | 14 +- .../clientpositive/union_remove_6_subq.q.out | 14 +- .../results/clientpositive/union_remove_7.q.out | 8 +- .../results/clientpositive/union_remove_8.q.out | 10 +- .../results/clientpositive/union_remove_9.q.out | 26 ++-- 179 files changed, 2340 insertions(+), 1405 deletions(-) diff --git a/ql/src/java/org/apache/hadoop/hive/ql/QueryProperties.java b/ql/src/java/org/apache/hadoop/hive/ql/QueryProperties.java index e8f7fba..3bc9432 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/QueryProperties.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/QueryProperties.java @@ -39,7 +39,6 @@ boolean noScanAnalyzeCommand; boolean analyzeRewrite; boolean ctas; - boolean insertToTable; int outerQueryLimit; boolean hasJoin = false; @@ -115,14 +114,6 @@ public void setCTAS(boolean ctas) { this.ctas = ctas; } - public boolean isInsertToTable() { - return insertToTable; - } - - public void setInsertToTable(boolean insertToTable) { - this.insertToTable = insertToTable; - } - public int getOuterQueryLimit() { return outerQueryLimit; } @@ -276,7 +267,6 @@ public void clear() { noScanAnalyzeCommand = false; analyzeRewrite = false; ctas = false; - insertToTable = false; outerQueryLimit = -1; hasJoin = false; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/StatsTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/StatsTask.java index 2a8167a..37611a9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/StatsTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/StatsTask.java @@ -174,13 +174,10 @@ private int aggregateStats() { if (statsAggregator != null) { String prefix = getAggregationPrefix(taskIndependent, table, null); updateStats(statsAggregator, parameters, prefix, maxPrefixLength, atomic); + parameters.put(StatsSetupConst.STATS_GENERATED_VIA_STATS_TASK, StatsSetupConst.TRUE); } updateQuickStats(wh, parameters, tTable.getSd()); - - // write table stats to metastore - parameters.put(StatsSetupConst.STATS_GENERATED_VIA_STATS_TASK, StatsSetupConst.TRUE); - db.alterTable(tableFullName, new Table(tTable)); console.printInfo("Table " + tableFullName + " stats: [" + toString(parameters) + ']'); @@ -333,8 +330,9 @@ private void updateQuickStats(Warehouse wh, Map parameters, private void clearStats(Map parameters) { for (String statType : StatsSetupConst.supportedStats) { if (parameters.containsKey(statType)) { - parameters.put(statType, "0"); + parameters.put(statType, "-1"); } + parameters.put(StatsSetupConst.COLUMN_STATS_ACCURATE, "false"); } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRFileSink1.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRFileSink1.java index e5b9c2b..588909b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRFileSink1.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRFileSink1.java @@ -61,6 +61,7 @@ public GenMRFileSink1() { * @param opProcCtx * context */ + @Override public Object process(Node nd, Stack stack, NodeProcessorCtx opProcCtx, Object... nodeOutputs) throws SemanticException { GenMRProcContext ctx = (GenMRProcContext) opProcCtx; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java index acd9bf5..477a90f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java @@ -983,7 +983,7 @@ public static TableScanOperator createTemporaryFile( fileSinkOp.setParentOperators(Utilities.makeList(parent)); // Create a dummy TableScanOperator for the file generated through fileSinkOp - TableScanOperator tableScanOp = (TableScanOperator) createTemporaryTableScanOperator( + TableScanOperator tableScanOp = createTemporaryTableScanOperator( parent.getSchema()); // Connect this TableScanOperator to child. @@ -1419,37 +1419,42 @@ public static void addStatsTask(FileSinkOperator nd, MoveTask mvTask, assert statsWork != null : "Error when genereting StatsTask"; statsWork.setSourceTask(currTask); - statsWork.setStatsReliable(hconf.getBoolVar(ConfVars.HIVE_STATS_RELIABLE)); - - if (currTask.getWork() instanceof MapredWork) { - MapredWork mrWork = (MapredWork) currTask.getWork(); - mrWork.getMapWork().setGatheringStats(true); - if (mrWork.getReduceWork() != null) { - mrWork.getReduceWork().setGatheringStats(true); - } - } else if (currTask.getWork() instanceof SparkWork) { - SparkWork work = (SparkWork) currTask.getWork(); - for (BaseWork w: work.getAllWork()) { - w.setGatheringStats(true); - } - } else { // must be TezWork - TezWork work = (TezWork) currTask.getWork(); - for (BaseWork w: work.getAllWork()) { - w.setGatheringStats(true); + if (hconf.getBoolVar(ConfVars.HIVESTATSAUTOGATHER)) { + statsWork.setStatsReliable(hconf.getBoolVar(ConfVars.HIVE_STATS_RELIABLE)); + + if (currTask.getWork() instanceof MapredWork) { + MapredWork mrWork = (MapredWork) currTask.getWork(); + mrWork.getMapWork().setGatheringStats(true); + if (mrWork.getReduceWork() != null) { + mrWork.getReduceWork().setGatheringStats(true); + } + } else if (currTask.getWork() instanceof SparkWork) { + SparkWork work = (SparkWork) currTask.getWork(); + for (BaseWork w: work.getAllWork()) { + w.setGatheringStats(true); + } + } else { // must be TezWork + TezWork work = (TezWork) currTask.getWork(); + for (BaseWork w: work.getAllWork()) { + w.setGatheringStats(true); + } } - } - // AggKey in StatsWork is used for stats aggregation while StatsAggPrefix - // in FileSinkDesc is used for stats publishing. They should be consistent. - statsWork.setAggKey(nd.getConf().getStatsAggPrefix()); - Task statsTask = TaskFactory.get(statsWork, hconf); + // AggKey in StatsWork is used for stats aggregation while StatsAggPrefix + // in FileSinkDesc is used for stats publishing. They should be consistent. + statsWork.setAggKey(nd.getConf().getStatsAggPrefix()); - // mark the MapredWork and FileSinkOperator for gathering stats - nd.getConf().setGatherStats(true); - nd.getConf().setStatsReliable(hconf.getBoolVar(ConfVars.HIVE_STATS_RELIABLE)); - nd.getConf().setMaxStatsKeyPrefixLength(StatsFactory.getMaxPrefixLength(hconf)); - // mrWork.addDestinationTable(nd.getConf().getTableInfo().getTableName()); + // mark the MapredWork and FileSinkOperator for gathering stats + nd.getConf().setGatherStats(true); + nd.getConf().setStatsReliable(hconf.getBoolVar(ConfVars.HIVE_STATS_RELIABLE)); + nd.getConf().setMaxStatsKeyPrefixLength(StatsFactory.getMaxPrefixLength(hconf)); + } else { + statsWork.setClearAggregatorStats(true); + statsWork.setNoStatsAggregator(true); + statsWork.setStatsReliable(false); + } + Task statsTask = TaskFactory.get(statsWork, hconf); // subscribe feeds from the MoveTask so that MoveTask can forward the list // of dynamic partition list to the StatsTask mvTask.addDependentTask(statsTask); @@ -1464,8 +1469,7 @@ public static void addStatsTask(FileSinkOperator nd, MoveTask mvTask, * @return */ public static boolean isInsertInto(ParseContext parseCtx, FileSinkOperator fsOp) { - return fsOp.getConf().getTableInfo().getTableName() != null && - parseCtx.getQueryProperties().isInsertToTable(); + return fsOp.getConf().getTableInfo().getTableName() != null; } /** @@ -1687,7 +1691,7 @@ public static boolean isMergeRequired(List> mvTasks, HiveConf hco // no need of merging if the move is to a local file system MoveTask mvTask = (MoveTask) GenMapRedUtils.findMoveTask(mvTasks, fsOp); - if (mvTask != null && isInsertTable && hconf.getBoolVar(ConfVars.HIVESTATSAUTOGATHER)) { + if (mvTask != null && isInsertTable) { GenMapRedUtils.addStatsTask(fsOp, mvTask, currTask, hconf); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/QBParseInfo.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/QBParseInfo.java index 14a7e9c..9072d7f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/QBParseInfo.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/QBParseInfo.java @@ -65,7 +65,6 @@ private final HashSet insertIntoTables; private boolean isAnalyzeCommand; // used for the analyze command (statistics) - private boolean isInsertToTable; // used for insert overwrite command (statistics) private boolean isNoScanAnalyzeCommand; // used for the analyze command (statistics) (noscan) private boolean isPartialScanAnalyzeCommand; // used for the analyze command (statistics) // (partialscan) @@ -550,14 +549,6 @@ public boolean isAnalyzeCommand() { return isAnalyzeCommand; } - public void setIsInsertToTable(boolean isInsertToTable) { - this.isInsertToTable = isInsertToTable; - } - - public boolean isInsertToTable() { - return isInsertToTable; - } - public void addTableSpec(String tName, TableSpec tSpec) { tableSpecs.put(tName, tSpec); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index 7f355e5..dbac0b2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -730,7 +730,7 @@ private ASTNode genValuesTempTable(ASTNode originalFrom, QB qb) throws SemanticE Path dataDir = null; if(!qb.getEncryptedTargetTablePaths().isEmpty()) { //currently only Insert into T values(...) is supported thus only 1 values clause - //and only 1 target table are possible. If/when support for + //and only 1 target table are possible. If/when support for //select ... from values(...) is added an insert statement may have multiple //encrypted target tables. dataDir = ctx.getMRTmpPath(qb.getEncryptedTargetTablePaths().get(0).toUri()); @@ -1761,8 +1761,6 @@ public void getMetaData(QB qb, ReadEntity parentInput) throws SemanticException qb.getMetaData().setDestForAlias(name, ts.partHandle); } if (HiveConf.getBoolVar(conf, HiveConf.ConfVars.HIVESTATSAUTOGATHER)) { - // Set that variable to automatically collect stats during the MapReduce job - qb.getParseInfo().setIsInsertToTable(true); // Add the table spec for the destination table. qb.getParseInfo().addTableSpec(ts.tableName.toLowerCase(), ts); } @@ -1799,8 +1797,6 @@ public void getMetaData(QB qb, ReadEntity parentInput) throws SemanticException } if (HiveConf.getBoolVar(conf, HiveConf.ConfVars.HIVESTATSAUTOGATHER)) { TableSpec ts = new TableSpec(db, conf, this.ast); - // Set that variable to automatically collect stats during the MapReduce job - qb.getParseInfo().setIsInsertToTable(true); // Add the table spec for the destination table. qb.getParseInfo().addTableSpec(ts.tableName.toLowerCase(), ts); } @@ -6379,7 +6375,6 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) // verify that our destination is empty before proceeding if (dest_tab.isImmutable() && qb.getParseInfo().isInsertIntoTable(dest_tab.getDbName(),dest_tab.getTableName())){ - qb.getParseInfo().isInsertToTable(); try { FileSystem fs = partPath.getFileSystem(conf); if (! MetaStoreUtils.isDirEmpty(fs,partPath)){ @@ -10070,7 +10065,7 @@ private void preProcessForInsert(ASTNode node, QB qb) throws SemanticException { return; } for (Node child : node.getChildren()) { - //each insert of multi insert looks like + //each insert of multi insert looks like //(TOK_INSERT (TOK_INSERT_INTO (TOK_TAB (TOK_TABNAME T1))) if (((ASTNode) child).getToken().getType() != HiveParser.TOK_INSERT) { continue; @@ -12199,7 +12194,6 @@ private void copyInfoToQueryProperties(QueryProperties queryProperties) { queryProperties.setNoScanAnalyzeCommand(qb.getParseInfo().isNoScanAnalyzeCommand()); queryProperties.setAnalyzeRewrite(qb.isAnalyzeRewrite()); queryProperties.setCTAS(qb.getTableDesc() != null); - queryProperties.setInsertToTable(qb.getParseInfo().isInsertToTable()); queryProperties.setHasOuterOrderBy(!qb.getParseInfo().getIsSubQ() && !qb.getParseInfo().getDestToOrderBy().isEmpty()); queryProperties.setOuterQueryLimit(qb.getParseInfo().getOuterQueryLimit()); diff --git a/ql/src/test/queries/clientpositive/insert_into1.q b/ql/src/test/queries/clientpositive/insert_into1.q index f19506a..311963d 100644 --- a/ql/src/test/queries/clientpositive/insert_into1.q +++ b/ql/src/test/queries/clientpositive/insert_into1.q @@ -31,6 +31,16 @@ explain SELECT COUNT(*) FROM insert_into1; select count(*) from insert_into1; -DROP TABLE insert_into1; +set hive.stats.autogather=false; +explain +insert into table insert_into1 values(1, 'abc'); +insert into table insert_into1 values(1, 'abc'); +explain +SELECT COUNT(*) FROM insert_into1; +select count(*) from insert_into1; + + +DROP TABLE insert_into1; +set hive.stats.autogather=true; set hive.compute.query.using.stats=false; diff --git a/ql/src/test/results/clientnegative/stats_partialscan_autogether.q.out b/ql/src/test/results/clientnegative/stats_partialscan_autogether.q.out index 321ebe5..03f9e9d 100644 --- a/ql/src/test/results/clientnegative/stats_partialscan_autogether.q.out +++ b/ql/src/test/results/clientnegative/stats_partialscan_autogether.q.out @@ -68,7 +68,7 @@ Table: analyze_srcpart_partial_scan Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/auto_join_filters.q.out b/ql/src/test/results/clientpositive/auto_join_filters.q.out index a6720d9..a8c0221 100644 --- a/ql/src/test/results/clientpositive/auto_join_filters.q.out +++ b/ql/src/test/results/clientpositive/auto_join_filters.q.out @@ -44,7 +44,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 3080335 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -330,7 +330,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 3078400 -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[10][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 diff --git a/ql/src/test/results/clientpositive/auto_join_nulls.q.out b/ql/src/test/results/clientpositive/auto_join_nulls.q.out index 4416f3e..7abde8f 100644 --- a/ql/src/test/results/clientpositive/auto_join_nulls.q.out +++ b/ql/src/test/results/clientpositive/auto_join_nulls.q.out @@ -14,7 +14,7 @@ POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/in1.txt' INTO TABLE my POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@myinput1 -Warning: Map Join MAPJOIN[18][bigTable=?] in task 'Stage-2:MAPRED' is a cross product +Warning: Map Join MAPJOIN[13][bigTable=b] in task 'Stage-2:MAPRED' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -24,7 +24,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 13630578 -Warning: Map Join MAPJOIN[17][bigTable=?] in task 'Stage-2:MAPRED' is a cross product +Warning: Map Join MAPJOIN[13][bigTable=a] in task 'Stage-2:MAPRED' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -34,7 +34,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 13630578 -Warning: Map Join MAPJOIN[17][bigTable=?] in task 'Stage-2:MAPRED' is a cross product +Warning: Map Join MAPJOIN[13][bigTable=b] in task 'Stage-2:MAPRED' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_1.q.out index 5114038..84d7fe4 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_1.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_1.q.out @@ -194,10 +194,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -242,10 +242,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -415,10 +415,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -463,10 +463,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -618,10 +618,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -714,10 +714,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -762,10 +762,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -809,10 +809,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -893,10 +893,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -940,10 +940,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1036,10 +1036,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1084,10 +1084,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1131,10 +1131,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1239,10 +1239,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1287,10 +1287,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out index e6e7ef3..6678fbb 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out @@ -166,10 +166,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -264,10 +264,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -311,10 +311,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -358,10 +358,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -516,10 +516,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -614,10 +614,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -661,10 +661,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -708,10 +708,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -864,10 +864,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -970,10 +970,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1017,10 +1017,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1188,10 +1188,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1237,10 +1237,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1282,10 +1282,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1407,10 +1407,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1454,10 +1454,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out index b2e782f..14d3f58 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out @@ -232,10 +232,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -283,10 +283,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -330,10 +330,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -381,10 +381,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -521,10 +521,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -569,10 +569,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -617,10 +617,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -665,10 +665,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out index 210f1ab..4cfa88a 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out @@ -174,10 +174,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -222,10 +222,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -379,10 +379,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -475,10 +475,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -523,10 +523,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -570,10 +570,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -654,10 +654,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -701,10 +701,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -797,10 +797,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -845,10 +845,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -892,10 +892,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1000,10 +1000,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1048,10 +1048,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out index a307b13..c0f270e 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out @@ -174,10 +174,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -346,10 +346,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -500,10 +500,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -546,10 +546,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -642,10 +642,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -689,10 +689,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -736,10 +736,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -819,10 +819,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -915,10 +915,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -962,10 +962,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1009,10 +1009,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1118,10 +1118,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out index f4ceee7..3142b37 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out @@ -190,10 +190,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -362,10 +362,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -516,10 +516,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -562,10 +562,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -658,10 +658,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -705,10 +705,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -752,10 +752,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -835,10 +835,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -931,10 +931,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -978,10 +978,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1025,10 +1025,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1134,10 +1134,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out index 3c2951a..22a0ac1 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out @@ -144,7 +144,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -154,6 +154,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -164,7 +166,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -174,6 +176,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -306,7 +310,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -316,6 +320,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -326,7 +332,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -336,6 +342,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -499,7 +507,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -509,6 +517,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -519,7 +529,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -529,6 +539,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -557,7 +569,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -567,6 +579,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -664,7 +678,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -674,6 +688,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -684,7 +700,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -694,6 +710,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -722,7 +740,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -732,6 +750,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -806,7 +826,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -816,6 +836,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -826,7 +848,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -836,6 +858,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out index e1f3888..3086d30 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out @@ -207,10 +207,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -255,10 +255,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -430,10 +430,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -478,10 +478,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -635,10 +635,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -681,10 +681,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -777,10 +777,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -825,10 +825,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -872,10 +872,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -919,10 +919,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1003,10 +1003,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1050,10 +1050,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1146,10 +1146,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1194,10 +1194,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1241,10 +1241,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1288,10 +1288,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1397,10 +1397,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1445,10 +1445,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_8.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_8.q.out index 38ecdbe..6528462 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_8.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_8.q.out @@ -207,10 +207,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -255,10 +255,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -430,10 +430,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -478,10 +478,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -637,10 +637,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -683,10 +683,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -779,10 +779,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -827,10 +827,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -874,10 +874,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -921,10 +921,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1005,10 +1005,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1052,10 +1052,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1148,10 +1148,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1196,10 +1196,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1243,10 +1243,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1290,10 +1290,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1399,10 +1399,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1447,10 +1447,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucket_map_join_1.q.out b/ql/src/test/results/clientpositive/bucket_map_join_1.q.out index 42e6a3f..3a00b88 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_1.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_1.q.out @@ -172,7 +172,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -182,8 +182,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -194,7 +194,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -204,8 +204,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucket_map_join_2.q.out b/ql/src/test/results/clientpositive/bucket_map_join_2.q.out index af73309..b07539c 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_2.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_2.q.out @@ -172,7 +172,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -182,8 +182,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -194,7 +194,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -204,8 +204,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out index 870ecdd..c5d7ceb 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out @@ -198,10 +198,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -316,10 +316,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -363,10 +363,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -561,10 +561,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -684,10 +684,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -731,10 +731,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out index 33f5c46..7ce3ea8 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out @@ -182,10 +182,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -300,10 +300,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -347,10 +347,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -545,10 +545,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -668,10 +668,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -715,10 +715,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out index 067d1ff..df6b016 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out @@ -182,10 +182,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -300,10 +300,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -347,10 +347,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -545,10 +545,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -668,10 +668,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -715,10 +715,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_1.q.out b/ql/src/test/results/clientpositive/bucketcontext_1.q.out index 77bfcf9..aaf1bd5 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_1.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_1.q.out @@ -168,10 +168,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -275,10 +275,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -323,10 +323,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -503,10 +503,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -551,10 +551,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_2.q.out b/ql/src/test/results/clientpositive/bucketcontext_2.q.out index a9db13d..a3420be 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_2.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_2.q.out @@ -152,10 +152,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -259,10 +259,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -307,10 +307,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -487,10 +487,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -535,10 +535,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_3.q.out b/ql/src/test/results/clientpositive/bucketcontext_3.q.out index 9ba3e0c..fd6e51f 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_3.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_3.q.out @@ -152,10 +152,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -198,10 +198,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -305,10 +305,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -484,10 +484,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_4.q.out b/ql/src/test/results/clientpositive/bucketcontext_4.q.out index a2b37a8..94a8042 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_4.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_4.q.out @@ -168,10 +168,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -214,10 +214,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -321,10 +321,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -500,10 +500,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_5.q.out b/ql/src/test/results/clientpositive/bucketcontext_5.q.out index 3ee1f0e..385410e 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_5.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_5.q.out @@ -182,7 +182,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -192,6 +192,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -202,7 +204,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -212,6 +214,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -351,7 +355,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -361,6 +365,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -371,7 +377,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -381,6 +387,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_6.q.out b/ql/src/test/results/clientpositive/bucketcontext_6.q.out index d2304fa..720a888 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_6.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_6.q.out @@ -211,10 +211,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -259,10 +259,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -437,10 +437,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -485,10 +485,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_7.q.out b/ql/src/test/results/clientpositive/bucketcontext_7.q.out index 1a105ed..bac65f2 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_7.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_7.q.out @@ -185,10 +185,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -231,10 +231,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -338,10 +338,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -386,10 +386,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -568,10 +568,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -616,10 +616,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketcontext_8.q.out b/ql/src/test/results/clientpositive/bucketcontext_8.q.out index 138e415..e7dbdd8 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_8.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_8.q.out @@ -185,10 +185,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -231,10 +231,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -338,10 +338,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -386,10 +386,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -568,10 +568,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -616,10 +616,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin1.q.out b/ql/src/test/results/clientpositive/bucketmapjoin1.q.out index 471ff73..7e83fba 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin1.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin1.q.out @@ -553,10 +553,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -668,7 +668,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -677,6 +677,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -687,7 +689,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -696,6 +698,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1175,10 +1179,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin10.q.out b/ql/src/test/results/clientpositive/bucketmapjoin10.q.out index b0e849d..6984683 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin10.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin10.q.out @@ -219,10 +219,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 3 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -264,10 +264,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -362,10 +362,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -409,10 +409,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 3 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin11.q.out b/ql/src/test/results/clientpositive/bucketmapjoin11.q.out index 4263cab..5b1b83b 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin11.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin11.q.out @@ -229,10 +229,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -274,10 +274,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -380,10 +380,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -427,10 +427,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -617,10 +617,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -662,10 +662,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -768,10 +768,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -815,10 +815,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin12.q.out b/ql/src/test/results/clientpositive/bucketmapjoin12.q.out index bcd7394..e170b44 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin12.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin12.q.out @@ -188,10 +188,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -293,10 +293,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -471,10 +471,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_3 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -569,10 +569,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin2.q.out b/ql/src/test/results/clientpositive/bucketmapjoin2.q.out index a8d9e9d..0c23c08 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin2.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin2.q.out @@ -191,10 +191,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -317,10 +317,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -737,10 +737,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -868,10 +868,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1335,10 +1335,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1380,10 +1380,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1511,10 +1511,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin3.q.out b/ql/src/test/results/clientpositive/bucketmapjoin3.q.out index c759f05..f5aba5a 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin3.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin3.q.out @@ -222,10 +222,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -348,10 +348,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -775,10 +775,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -906,10 +906,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin4.q.out b/ql/src/test/results/clientpositive/bucketmapjoin4.q.out index f61500c..ee50787 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin4.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin4.q.out @@ -277,7 +277,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -286,6 +286,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -296,7 +298,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -305,6 +307,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -758,7 +762,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -767,6 +771,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -777,7 +783,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -786,6 +792,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin5.q.out b/ql/src/test/results/clientpositive/bucketmapjoin5.q.out index 0cb2825..67cf64c 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin5.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin5.q.out @@ -338,10 +338,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -385,10 +385,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -884,10 +884,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -931,10 +931,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin7.q.out b/ql/src/test/results/clientpositive/bucketmapjoin7.q.out index 667a9db..15b0057 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin7.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin7.q.out @@ -168,10 +168,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -275,10 +275,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin8.q.out b/ql/src/test/results/clientpositive/bucketmapjoin8.q.out index 252b377..e20abd5 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin8.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin8.q.out @@ -153,10 +153,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -259,10 +259,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -446,10 +446,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -552,10 +552,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin9.q.out b/ql/src/test/results/clientpositive/bucketmapjoin9.q.out index 5e28dc3..c4da41f 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin9.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin9.q.out @@ -161,10 +161,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 3 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -259,10 +259,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -479,10 +479,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -577,10 +577,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out index 6ae127d..40c5d46 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out @@ -166,10 +166,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -273,7 +273,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -282,6 +282,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -292,7 +294,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -301,6 +303,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out index 4c9f54a..c4f43c3 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out @@ -168,10 +168,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -213,10 +213,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -328,7 +328,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -337,6 +337,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -347,7 +349,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -356,6 +358,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out index 9a0bfc4..b8e4dab 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out @@ -286,7 +286,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -296,6 +296,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -306,7 +308,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -316,6 +318,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -475,7 +479,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -485,6 +489,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -495,7 +501,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -505,6 +511,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -653,7 +661,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -663,6 +671,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -673,7 +683,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -683,6 +693,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -834,7 +846,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -844,6 +856,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -854,7 +868,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -864,6 +878,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1015,7 +1031,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1025,6 +1041,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1035,7 +1053,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1045,6 +1063,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1196,7 +1216,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1206,6 +1226,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1216,7 +1238,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1226,6 +1248,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1377,7 +1401,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1387,6 +1411,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1397,7 +1423,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1407,6 +1433,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1558,7 +1586,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1568,6 +1596,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1578,7 +1608,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1588,6 +1618,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1739,7 +1771,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1749,6 +1781,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1759,7 +1793,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1769,6 +1803,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/columnstats_partlvl.q.out b/ql/src/test/results/clientpositive/columnstats_partlvl.q.out index e0c4cfe..c2eeada 100644 --- a/ql/src/test/results/clientpositive/columnstats_partlvl.q.out +++ b/ql/src/test/results/clientpositive/columnstats_partlvl.q.out @@ -149,10 +149,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.employee_part numFiles 1 - numRows 0 + numRows -1 partition_columns employeesalary partition_columns.types double - rawDataSize 0 + rawDataSize -1 serialization.ddl struct employee_part { i32 employeeid, string employeename} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -349,10 +349,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.employee_part numFiles 1 - numRows 0 + numRows -1 partition_columns employeesalary partition_columns.types double - rawDataSize 0 + rawDataSize -1 serialization.ddl struct employee_part { i32 employeeid, string employeename} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out b/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out index 19283bb..58c9efd 100644 --- a/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out +++ b/ql/src/test/results/clientpositive/columnstats_tbllvl.q.out @@ -134,7 +134,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite columns.comments @@ -143,6 +143,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -153,7 +155,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite columns.comments @@ -162,6 +164,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -536,7 +540,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite columns.comments @@ -545,6 +549,8 @@ STAGE PLANS: #### A masked pattern was here #### name dummydb.uservisits_in_dummy_db numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_in_dummy_db { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -555,7 +561,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite columns.comments @@ -564,6 +570,8 @@ STAGE PLANS: #### A masked pattern was here #### name dummydb.uservisits_in_dummy_db numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_in_dummy_db { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out b/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out index 7c91248..12c5b77 100644 --- a/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out +++ b/ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out @@ -150,7 +150,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite columns.comments @@ -159,6 +159,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -169,7 +171,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite columns.comments @@ -178,6 +180,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out index 939e206..8c2ce90 100644 --- a/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out +++ b/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_dynamic.q.out @@ -301,7 +301,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns key,value columns.comments 'default','default' @@ -309,8 +309,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.src numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -321,7 +321,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns key,value columns.comments 'default','default' @@ -329,8 +329,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.src numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out index fd7932e..ddc1346 100644 --- a/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out +++ b/ql/src/test/results/clientpositive/encrypted/encryption_insert_partition_static.q.out @@ -307,7 +307,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns key,value columns.comments 'default','default' @@ -315,8 +315,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.src numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -327,7 +327,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns key,value columns.comments 'default','default' @@ -335,8 +335,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.src numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out index 9b6f750..631812a 100644 --- a/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out +++ b/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out @@ -665,7 +665,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns key,value columns.comments 'default','default' @@ -673,8 +673,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.src numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -685,7 +685,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count -1 columns key,value columns.comments 'default','default' @@ -693,8 +693,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.src numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/groupby_sort_6.q.out b/ql/src/test/results/clientpositive/groupby_sort_6.q.out index 0169430..95fb19e 100644 --- a/ql/src/test/results/clientpositive/groupby_sort_6.q.out +++ b/ql/src/test/results/clientpositive/groupby_sort_6.q.out @@ -449,10 +449,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.t1 numFiles 1 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct t1 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/insert_into1.q.out b/ql/src/test/results/clientpositive/insert_into1.q.out index 9e5f3bb..4115e88 100644 --- a/ql/src/test/results/clientpositive/insert_into1.q.out +++ b/ql/src/test/results/clientpositive/insert_into1.q.out @@ -337,6 +337,161 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 10 +PREHOOK: query: explain +insert into table insert_into1 values(1, 'abc') +PREHOOK: type: QUERY +POSTHOOK: query: explain +insert into table insert_into1 values(1, 'abc') +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 + Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + +STAGE PLANS: + Stage: Stage-1 + Map Reduce + Map Operator Tree: + TableScan + alias: values__tmp__table__1 + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(tmp_values_col1) (type: int), tmp_values_col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-7 + Conditional Operator + + Stage: Stage-4 + Move Operator + files: + hdfs directory: true +#### A masked pattern was here #### + + Stage: Stage-0 + Move Operator + tables: + replace: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 + Map Reduce + Map Operator Tree: + TableScan + File Output Operator + compressed: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-5 + Map Reduce + Map Operator Tree: + TableScan + File Output Operator + compressed: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-6 + Move Operator + files: + hdfs directory: true +#### A masked pattern was here #### + +PREHOOK: query: insert into table insert_into1 values(1, 'abc') +PREHOOK: type: QUERY +PREHOOK: Input: default@values__tmp__table__2 +PREHOOK: Output: default@insert_into1 +POSTHOOK: query: insert into table insert_into1 values(1, 'abc') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@values__tmp__table__2 +POSTHOOK: Output: default@insert_into1 +POSTHOOK: Lineage: insert_into1.key EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ] +POSTHOOK: Lineage: insert_into1.value SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ] +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Map Reduce + Map Operator Tree: + TableScan + alias: insert_into1 + Statistics: Num rows: -1 Data size: 120 Basic stats: PARTIAL Column stats: COMPLETE + Select Operator + Statistics: Num rows: -1 Data size: 120 Basic stats: PARTIAL Column stats: COMPLETE + Group By Operator + aggregations: count() + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +11 PREHOOK: query: DROP TABLE insert_into1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into1 diff --git a/ql/src/test/results/clientpositive/join_filters.q.out b/ql/src/test/results/clientpositive/join_filters.q.out index 4f112bd..8a4e2a2 100644 --- a/ql/src/test/results/clientpositive/join_filters.q.out +++ b/ql/src/test/results/clientpositive/join_filters.q.out @@ -57,7 +57,7 @@ NULL NULL 12 35 NULL NULL 48 NULL NULL NULL NULL 40 LNZKrcVNAvaeDALnsg72bw== -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -807,7 +807,7 @@ POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 100 100 100 100 YaI1msgLVpfEx943Tfea/Q== -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[10][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 diff --git a/ql/src/test/results/clientpositive/join_nulls.q.out b/ql/src/test/results/clientpositive/join_nulls.q.out index 46e0170..d5f20b7 100644 --- a/ql/src/test/results/clientpositive/join_nulls.q.out +++ b/ql/src/test/results/clientpositive/join_nulls.q.out @@ -18,7 +18,7 @@ POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/in1.txt' INTO TABLE my POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@myinput1 -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT * FROM myinput1 a JOIN myinput1 b PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -36,7 +36,7 @@ POSTHOOK: Input: default@myinput1 NULL 35 100 100 NULL 35 48 NULL NULL 35 NULL 35 -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -54,7 +54,7 @@ POSTHOOK: Input: default@myinput1 NULL 35 100 100 NULL 35 48 NULL NULL 35 NULL 35 -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product PREHOOK: query: SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 diff --git a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out index a9522e0..92fd44f 100644 --- a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out +++ b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out @@ -489,8 +489,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 3 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 10586 #### A masked pattern was here #### @@ -634,10 +634,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.list_bucketing_dynamic_part numFiles 3 - numRows 0 + numRows -1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe diff --git a/ql/src/test/results/clientpositive/parquet_serde.q.out b/ql/src/test/results/clientpositive/parquet_serde.q.out index e753180..dbd9e27 100644 --- a/ql/src/test/results/clientpositive/parquet_serde.q.out +++ b/ql/src/test/results/clientpositive/parquet_serde.q.out @@ -75,8 +75,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 36 #### A masked pattern was here #### @@ -182,8 +182,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 36 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx_cbo_2.q.out b/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx_cbo_2.q.out index 3ee2e0f..061324c 100644 --- a/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx_cbo_2.q.out +++ b/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx_cbo_2.q.out @@ -116,9 +116,9 @@ STAGE PLANS: Map Operator Tree: TableScan alias: lineitem_ix - Statistics: Num rows: 1 Data size: 12099 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: -1 Data size: 12099 Basic stats: PARTIAL Column stats: COMPLETE Select Operator - Statistics: Num rows: 1 Data size: 12099 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: -1 Data size: 12099 Basic stats: PARTIAL Column stats: COMPLETE Group By Operator aggregations: count(1) mode: hash @@ -3814,22 +3814,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: lineitem_ix - Statistics: Num rows: 1 Data size: 12099 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: -1 Data size: 12099 Basic stats: PARTIAL Column stats: COMPLETE Select Operator expressions: 1 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 12099 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: -1 Data size: 12099 Basic stats: PARTIAL Column stats: COMPLETE Group By Operator aggregations: count(_col0) keys: _col0 (type: int) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 9223372036854775807 Data size: 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 9223372036854775807 Data size: 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: bigint) Reduce Operator Tree: Group By Operator @@ -3837,10 +3837,10 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat diff --git a/ql/src/test/results/clientpositive/skewjoin_union_remove_1.q.out b/ql/src/test/results/clientpositive/skewjoin_union_remove_1.q.out index 1f21877..9ddccaf 100644 --- a/ql/src/test/results/clientpositive/skewjoin_union_remove_1.q.out +++ b/ql/src/test/results/clientpositive/skewjoin_union_remove_1.q.out @@ -336,8 +336,9 @@ SELECT * FROM T1 a JOIN T2 b ON a.key = b.key POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage + Stage-0 depends on stages: Stage-1, Stage-4 + Stage-2 depends on stages: Stage-0 + Stage-4 is a root stage STAGE PLANS: Stage: Stage-1 @@ -399,7 +400,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 - Stage: Stage-3 + Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-4 Map Reduce Map Operator Tree: TableScan @@ -487,8 +491,9 @@ SELECT * FROM T1 a RIGHT OUTER JOIN T2 b ON a.key = b.key POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage + Stage-0 depends on stages: Stage-1, Stage-4 + Stage-2 depends on stages: Stage-0 + Stage-4 is a root stage STAGE PLANS: Stage: Stage-1 @@ -550,7 +555,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 - Stage: Stage-3 + Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-4 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out index 09d2692..664c13d 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out @@ -202,10 +202,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -250,10 +250,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -434,10 +434,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -482,10 +482,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -657,10 +657,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -756,10 +756,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -804,10 +804,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out index a70b161..979aaf8 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out @@ -251,10 +251,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -337,10 +337,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -403,10 +403,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -506,10 +506,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -554,10 +554,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out index 2ea0a65..3c17fb9 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out @@ -182,10 +182,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -230,10 +230,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -407,10 +407,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -506,10 +506,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -554,10 +554,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out index 6281929..26da368 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out @@ -182,10 +182,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -365,10 +365,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -539,10 +539,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -587,10 +587,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -687,10 +687,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out index 31e9d86..be7d048 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out @@ -198,10 +198,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -381,10 +381,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -555,10 +555,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -603,10 +603,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -703,10 +703,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out index 3eceb0b..b37892c 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out @@ -152,7 +152,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -162,6 +162,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -172,7 +174,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -182,6 +184,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -325,7 +329,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -335,6 +339,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -345,7 +351,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -355,6 +361,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -489,7 +497,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -499,6 +507,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -509,7 +519,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -519,6 +529,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -584,7 +596,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -594,6 +606,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -604,7 +618,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -614,6 +628,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out index ddbca05..d82aef1 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out @@ -215,10 +215,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -263,10 +263,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -449,10 +449,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -497,10 +497,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -674,10 +674,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -722,10 +722,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -822,10 +822,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -870,10 +870,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out index 88d4dcb..df4aac1 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out @@ -215,10 +215,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -263,10 +263,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -449,10 +449,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -497,10 +497,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -676,10 +676,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -724,10 +724,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -824,10 +824,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -872,10 +872,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out index 7570ebe..ffe6b24 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out @@ -139,7 +139,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name value @@ -149,8 +149,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table2 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -161,7 +161,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name value @@ -171,8 +171,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table2 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -232,7 +232,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -242,8 +242,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -254,7 +254,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -264,8 +264,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out index 80b44e9..644d990 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out @@ -139,7 +139,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name value @@ -149,8 +149,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table2 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -161,7 +161,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name value @@ -171,8 +171,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table2 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -232,7 +232,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -242,8 +242,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -254,7 +254,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key @@ -264,8 +264,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.table1 numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out index 6230bef..39d077b 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark1.q.out @@ -221,10 +221,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -337,10 +337,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -553,10 +553,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -668,10 +668,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out index 1a33625..d59f2ea 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark2.q.out @@ -205,10 +205,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -321,10 +321,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -542,10 +542,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -663,10 +663,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out index fed923c..18a3865 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark3.q.out @@ -205,10 +205,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -321,10 +321,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -537,10 +537,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -652,10 +652,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out index d4a9c98..6617a0a 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out @@ -584,10 +584,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -689,7 +689,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -698,6 +698,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -708,7 +710,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -717,6 +719,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -966,7 +970,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -975,6 +979,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -985,7 +991,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -994,6 +1000,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1093,10 +1101,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out index 678ad54..3bb2526 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out @@ -238,10 +238,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 3 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -285,10 +285,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -378,10 +378,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -425,10 +425,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 3 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out index 95606f0..e0e9fea 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out @@ -253,10 +253,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -300,10 +300,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -399,10 +399,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -446,10 +446,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -661,10 +661,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -708,10 +708,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -807,10 +807,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -854,10 +854,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 4 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out index d6c25e4..1cb89be 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out @@ -212,10 +212,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -309,10 +309,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -507,10 +507,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_3 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -599,10 +599,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out index d82480e..5233710 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out @@ -210,10 +210,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -326,10 +326,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -612,10 +612,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -733,10 +733,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1036,10 +1036,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1083,10 +1083,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1205,10 +1205,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out index 39552c1..ba216e2 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out @@ -241,10 +241,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -357,10 +357,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -650,10 +650,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -771,10 +771,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out index ad2762d..a7aad4d 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out @@ -216,7 +216,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -225,6 +225,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -235,7 +237,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -244,6 +246,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -327,7 +331,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -336,6 +340,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -346,7 +352,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -355,6 +361,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -589,7 +597,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -598,6 +606,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -608,7 +618,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -617,6 +627,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -705,7 +717,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -714,6 +726,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -724,7 +738,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -733,6 +747,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out index f7c3d4d..bb5465a 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out @@ -266,7 +266,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -275,6 +275,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -285,7 +287,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -294,6 +296,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -388,10 +392,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -435,10 +439,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -704,7 +708,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -713,6 +717,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -723,7 +729,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -732,6 +738,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -831,10 +839,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -878,10 +886,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out index 7bfe440..3f5cc64 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out @@ -192,10 +192,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -291,10 +291,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out index 4601eb1..fa05322 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out @@ -177,10 +177,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -275,10 +275,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -487,10 +487,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -585,10 +585,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out index 60bd103..7e18e59 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out @@ -180,10 +180,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 3 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -272,10 +272,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -512,10 +512,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -604,10 +604,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out index 031c46c..7dbc55a 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out @@ -180,10 +180,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -279,7 +279,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -288,6 +288,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -298,7 +300,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -307,6 +309,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out index 4a8f46d..acb2f6f 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out @@ -187,10 +187,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -234,10 +234,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -340,7 +340,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -349,6 +349,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -359,7 +361,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -368,6 +370,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out index a09904e..9c72fb3 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out @@ -235,7 +235,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -245,6 +245,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -255,7 +257,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -265,6 +267,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -343,7 +347,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -353,6 +357,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -363,7 +369,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -373,6 +379,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -481,7 +489,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -491,6 +499,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -501,7 +511,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -511,6 +521,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -589,7 +601,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -599,6 +611,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -609,7 +623,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -619,6 +633,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -719,7 +735,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -729,6 +745,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -739,7 +757,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -749,6 +767,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -821,7 +841,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -831,6 +851,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -841,7 +863,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -851,6 +873,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -954,7 +978,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -964,6 +988,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -974,7 +1000,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -984,6 +1010,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1056,7 +1084,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1066,6 +1094,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1076,7 +1106,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1086,6 +1116,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1189,7 +1221,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1199,6 +1231,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1209,7 +1243,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1219,6 +1253,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1291,7 +1327,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1301,6 +1337,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1311,7 +1349,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1321,6 +1359,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1424,7 +1464,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1434,6 +1474,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test4 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1444,7 +1486,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1454,6 +1496,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test4 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1526,7 +1570,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1536,6 +1580,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1546,7 +1592,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1556,6 +1602,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test1 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1659,7 +1707,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1669,6 +1717,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1679,7 +1729,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -1689,6 +1739,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1761,7 +1813,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1771,6 +1823,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1781,7 +1835,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1791,6 +1845,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1894,7 +1950,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1904,6 +1960,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test4 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1914,7 +1972,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -1924,6 +1982,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test4 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1996,7 +2056,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -2006,6 +2066,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -2016,7 +2078,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -2026,6 +2088,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test2 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -2129,7 +2193,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -2139,6 +2203,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test4 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -2149,7 +2215,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name value @@ -2159,6 +2225,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test4 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -2231,7 +2299,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -2241,6 +2309,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -2251,7 +2321,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 3 bucket_field_name key @@ -2261,6 +2331,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.test3 numFiles 3 + numRows -1 + rawDataSize -1 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/spark/count.q.out b/ql/src/test/results/clientpositive/spark/count.q.out index cb9eda5..6923a5f 100644 --- a/ql/src/test/results/clientpositive/spark/count.q.out +++ b/ql/src/test/results/clientpositive/spark/count.q.out @@ -123,11 +123,11 @@ STAGE PLANS: Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) - outputColumnNames: _col1, _col2, _col3, _col4 + outputColumnNames: a, b, c, d Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Group By Operator - aggregations: count(1), count(), count(_col1), count(_col2), count(_col3), count(_col4), count(DISTINCT _col1), count(DISTINCT _col2), count(DISTINCT _col3), count(DISTINCT _col4), count(DISTINCT _col1, _col2), count(DISTINCT _col2, _col3), count(DISTINCT _col3, _col4), count(DISTINCT _col1, _col4), count(DISTINCT _col1, _col3), count(DISTINCT _col2, _col4), count(DISTINCT _col1, _col2, _col3), count(DISTINCT _col2, _col3, _col4), count(DISTINCT _col1, _col3, _col4), count(DISTINCT _col1, _col2, _col4), count(DISTINCT _col1, _col2, _col3, _col4) - keys: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int) + aggregations: count(1), count(), count(a), count(b), count(c), count(d), count(DISTINCT a), count(DISTINCT b), count(DISTINCT c), count(DISTINCT d), count(DISTINCT a, b), count(DISTINCT b, c), count(DISTINCT c, d), count(DISTINCT a, d), count(DISTINCT a, c), count(DISTINCT b, d), count(DISTINCT a, b, c), count(DISTINCT b, c, d), count(DISTINCT a, c, d), count(DISTINCT a, b, d), count(DISTINCT a, b, c, d) + keys: a (type: int), b (type: int), c (type: int), d (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24 Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE @@ -252,10 +252,10 @@ STAGE PLANS: Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) - outputColumnNames: _col1, _col2, _col3, _col4 + outputColumnNames: a, b, c, d Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int) + key expressions: a (type: int), b (type: int), c (type: int), d (type: int) sort order: ++++ Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 2 diff --git a/ql/src/test/results/clientpositive/spark/insert_into1.q.out b/ql/src/test/results/clientpositive/spark/insert_into1.q.out index e7f7d8f..7e41c06 100644 --- a/ql/src/test/results/clientpositive/spark/insert_into1.q.out +++ b/ql/src/test/results/clientpositive/spark/insert_into1.q.out @@ -355,6 +355,126 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 10 +PREHOOK: query: explain +insert into table insert_into1 values(1, 'abc') +PREHOOK: type: QUERY +POSTHOOK: query: explain +insert into table insert_into1 values(1, 'abc') +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Spark +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: values__tmp__table__1 + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(tmp_values_col1) (type: int), tmp_values_col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-0 + Move Operator + tables: + replace: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-2 + Stats-Aggr Operator + +PREHOOK: query: insert into table insert_into1 values(1, 'abc') +PREHOOK: type: QUERY +PREHOOK: Input: default@values__tmp__table__2 +PREHOOK: Output: default@insert_into1 +POSTHOOK: query: insert into table insert_into1 values(1, 'abc') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@values__tmp__table__2 +POSTHOOK: Output: default@insert_into1 +POSTHOOK: Lineage: insert_into1.key EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ] +POSTHOOK: Lineage: insert_into1.value SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ] +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Spark + Edges: + Reducer 2 <- Map 1 (GROUP, 1) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: insert_into1 + Statistics: Num rows: -1 Data size: 120 Basic stats: PARTIAL Column stats: COMPLETE + Select Operator + Statistics: Num rows: -1 Data size: 120 Basic stats: PARTIAL Column stats: COMPLETE + Group By Operator + aggregations: count() + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +11 PREHOOK: query: DROP TABLE insert_into1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into1 diff --git a/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out b/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out index f6acc34..564c81d 100644 --- a/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out +++ b/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out @@ -349,6 +349,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -463,6 +464,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: INSERT OVERWRITE TABLE DEST1 SELECT * FROM T1 a JOIN T2 b ON a.key = b.key PREHOOK: type: QUERY @@ -506,6 +510,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -620,6 +625,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: INSERT OVERWRITE TABLE DEST1 SELECT * FROM T1 a RIGHT OUTER JOIN T2 b ON a.key = b.key PREHOOK: type: QUERY diff --git a/ql/src/test/results/clientpositive/spark/stats1.q.out b/ql/src/test/results/clientpositive/spark/stats1.q.out index f00db10..6e27a9d 100644 --- a/ql/src/test/results/clientpositive/spark/stats1.q.out +++ b/ql/src/test/results/clientpositive/spark/stats1.q.out @@ -225,10 +225,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 3 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 1583 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/spark/stats12.q.out b/ql/src/test/results/clientpositive/spark/stats12.q.out index db575df..ef7f879 100644 --- a/ql/src/test/results/clientpositive/spark/stats12.q.out +++ b/ql/src/test/results/clientpositive/spark/stats12.q.out @@ -81,7 +81,7 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' @@ -127,7 +127,7 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' @@ -334,7 +334,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -376,7 +376,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/spark/stats13.q.out b/ql/src/test/results/clientpositive/spark/stats13.q.out index f38f876..aa45a62 100644 --- a/ql/src/test/results/clientpositive/spark/stats13.q.out +++ b/ql/src/test/results/clientpositive/spark/stats13.q.out @@ -82,7 +82,7 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' @@ -242,7 +242,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -284,7 +284,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -326,7 +326,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/spark/stats18.q.out b/ql/src/test/results/clientpositive/spark/stats18.q.out index a061846..64a765d 100644 --- a/ql/src/test/results/clientpositive/spark/stats18.q.out +++ b/ql/src/test/results/clientpositive/spark/stats18.q.out @@ -99,8 +99,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 7170 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/spark/stats2.q.out b/ql/src/test/results/clientpositive/spark/stats2.q.out index ddc8226..21e9970 100644 --- a/ql/src/test/results/clientpositive/spark/stats2.q.out +++ b/ql/src/test/results/clientpositive/spark/stats2.q.out @@ -15,6 +15,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -52,6 +53,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.analyze_t1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table analyze_t1 partition (ds, hr) select * from srcpart where ds is not null PREHOOK: type: QUERY PREHOOK: Input: default@srcpart diff --git a/ql/src/test/results/clientpositive/spark/stats3.q.out b/ql/src/test/results/clientpositive/spark/stats3.q.out index dd3a95b..23fdeca 100644 --- a/ql/src/test/results/clientpositive/spark/stats3.q.out +++ b/ql/src/test/results/clientpositive/spark/stats3.q.out @@ -87,10 +87,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 11 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/spark/stats6.q.out b/ql/src/test/results/clientpositive/spark/stats6.q.out index b4435f2..e6f4986 100644 --- a/ql/src/test/results/clientpositive/spark/stats6.q.out +++ b/ql/src/test/results/clientpositive/spark/stats6.q.out @@ -166,7 +166,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -208,7 +208,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/spark/stats_noscan_1.q.out b/ql/src/test/results/clientpositive/spark/stats_noscan_1.q.out index 80c3092..6637881 100644 --- a/ql/src/test/results/clientpositive/spark/stats_noscan_1.q.out +++ b/ql/src/test/results/clientpositive/spark/stats_noscan_1.q.out @@ -187,7 +187,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -229,7 +229,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -468,7 +468,7 @@ Table: analyze_srcpart_partial Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -510,7 +510,7 @@ Table: analyze_srcpart_partial Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/spark/stats_noscan_2.q.out b/ql/src/test/results/clientpositive/spark/stats_noscan_2.q.out index 238727d..9ac9bb5 100644 --- a/ql/src/test/results/clientpositive/spark/stats_noscan_2.q.out +++ b/ql/src/test/results/clientpositive/spark/stats_noscan_2.q.out @@ -52,7 +52,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: EXTERNAL_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false EXTERNAL TRUE numFiles 0 numRows -1 diff --git a/ql/src/test/results/clientpositive/spark/stats_partscan_1_23.q.out b/ql/src/test/results/clientpositive/spark/stats_partscan_1_23.q.out index adcf150..c4eff7a 100644 --- a/ql/src/test/results/clientpositive/spark/stats_partscan_1_23.q.out +++ b/ql/src/test/results/clientpositive/spark/stats_partscan_1_23.q.out @@ -78,7 +78,7 @@ Table: analyze_srcpart_partial_scan Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -153,7 +153,7 @@ Table: analyze_srcpart_partial_scan Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -195,7 +195,7 @@ Table: analyze_srcpart_partial_scan Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/spark/union_remove_1.q.out b/ql/src/test/results/clientpositive/spark/union_remove_1.q.out index bf0fc20..c598b57 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_1.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_1.q.out @@ -63,6 +63,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -157,6 +158,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_10.q.out b/ql/src/test/results/clientpositive/spark/union_remove_10.q.out index a9d4b51..cf56d4e 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_10.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_10.q.out @@ -78,12 +78,13 @@ select * FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -164,10 +165,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -184,6 +185,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -194,7 +198,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -205,7 +209,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_11.q.out b/ql/src/test/results/clientpositive/spark/union_remove_11.q.out index be65741..f6ec81c 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_11.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_11.q.out @@ -78,12 +78,13 @@ select * FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -154,10 +155,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -174,6 +175,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -184,7 +188,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -195,7 +199,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_12.q.out b/ql/src/test/results/clientpositive/spark/union_remove_12.q.out index 62d9d7d..0a4d84d 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_12.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_12.q.out @@ -69,17 +69,18 @@ FROM inputTbl1 a join inputTbl1 b on a.key=b.key )c POSTHOOK: type: QUERY STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-1 depends on stages: Stage-7 - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-8 is a root stage + Stage-1 depends on stages: Stage-8 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: - Stage: Stage-7 + Stage: Stage-8 Spark #### A masked pattern was here #### Vertices: @@ -160,10 +161,10 @@ STAGE PLANS: Local Work: Map Reduce Local Work - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -180,6 +181,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -190,7 +194,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -201,7 +205,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_13.q.out b/ql/src/test/results/clientpositive/spark/union_remove_13.q.out index 995d180..317d621 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_13.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_13.q.out @@ -69,17 +69,18 @@ FROM inputTbl1 a join inputTbl1 b on a.key=b.key )c POSTHOOK: type: QUERY STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-1 depends on stages: Stage-7 - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-8 is a root stage + Stage-1 depends on stages: Stage-8 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: - Stage: Stage-7 + Stage: Stage-8 Spark #### A masked pattern was here #### Vertices: @@ -186,10 +187,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -206,6 +207,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -216,7 +220,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -227,7 +231,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_14.q.out b/ql/src/test/results/clientpositive/spark/union_remove_14.q.out index b28bb75..442516e 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_14.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_14.q.out @@ -71,17 +71,18 @@ FROM inputTbl1 a join inputTbl1 b on a.key=b.key )c POSTHOOK: type: QUERY STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-1 depends on stages: Stage-7 - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-8 is a root stage + Stage-1 depends on stages: Stage-8 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: - Stage: Stage-7 + Stage: Stage-8 Spark #### A masked pattern was here #### Vertices: @@ -162,10 +163,10 @@ STAGE PLANS: Local Work: Map Reduce Local Work - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -182,6 +183,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -192,7 +196,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -203,7 +207,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_15.q.out b/ql/src/test/results/clientpositive/spark/union_remove_15.q.out index 58abd21..adf0df0 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_15.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_15.q.out @@ -69,6 +69,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -173,6 +174,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_16.q.out b/ql/src/test/results/clientpositive/spark/union_remove_16.q.out index 356e79a..7fa371c 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_16.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_16.q.out @@ -66,12 +66,13 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -164,10 +165,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -186,6 +187,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -196,7 +200,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -207,7 +211,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_17.q.out b/ql/src/test/results/clientpositive/spark/union_remove_17.q.out index 74a5b23..1dfe70c 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_17.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_17.q.out @@ -63,6 +63,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -124,6 +125,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_18.q.out b/ql/src/test/results/clientpositive/spark/union_remove_18.q.out index 2c01a5c..306fbc7 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_18.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_18.q.out @@ -67,6 +67,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -171,6 +172,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_19.q.out b/ql/src/test/results/clientpositive/spark/union_remove_19.q.out index d1e2312..3095f17 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_19.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_19.q.out @@ -67,6 +67,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -161,6 +162,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT a.key, a.`values` FROM ( @@ -260,6 +264,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -376,6 +381,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT a.key, a.`values` FROM ( @@ -437,6 +445,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -545,6 +554,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 select key, `values` from ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_2.q.out b/ql/src/test/results/clientpositive/spark/union_remove_2.q.out index 59d88cb..524a8c9 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_2.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_2.q.out @@ -69,6 +69,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -159,6 +160,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_20.q.out b/ql/src/test/results/clientpositive/spark/union_remove_20.q.out index f57353f..72dd945 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_20.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_20.q.out @@ -65,6 +65,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -167,6 +168,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT a.`values`, a.key FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_21.q.out b/ql/src/test/results/clientpositive/spark/union_remove_21.q.out index 48867fb..66c51f1 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_21.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_21.q.out @@ -65,6 +65,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -175,6 +176,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT a.key FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_22.q.out b/ql/src/test/results/clientpositive/spark/union_remove_22.q.out index c41e12f..a9b6d9b 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_22.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_22.q.out @@ -67,6 +67,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -169,6 +170,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT a.key, a.`values`, a.`values` FROM ( @@ -268,6 +272,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -370,6 +375,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT a.key, concat(a.`values`, a.`values`), concat(a.`values`, a.`values`) FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_23.q.out b/ql/src/test/results/clientpositive/spark/union_remove_23.q.out index d9875f3..fb828b6 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_23.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_23.q.out @@ -67,6 +67,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -189,6 +190,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_24.q.out b/ql/src/test/results/clientpositive/spark/union_remove_24.q.out index e989ed2..da939e6 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_24.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_24.q.out @@ -61,6 +61,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -163,6 +164,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_25.q.out b/ql/src/test/results/clientpositive/spark/union_remove_25.q.out index a5a0126..9892eea 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_25.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_25.q.out @@ -79,6 +79,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -175,6 +176,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 partition(ds='2004') SELECT * FROM ( @@ -221,7 +225,7 @@ Table: outputtbl1 Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 4 numRows -1 rawDataSize -1 @@ -279,6 +283,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -375,6 +380,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl2 partition(ds) SELECT * FROM ( @@ -432,7 +440,7 @@ Table: outputtbl2 Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 2 numRows -1 rawDataSize -1 @@ -468,6 +476,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -565,6 +574,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl3 partition(ds, hr) SELECT * FROM ( @@ -627,7 +639,7 @@ Table: outputtbl3 Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 2 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/spark/union_remove_3.q.out b/ql/src/test/results/clientpositive/spark/union_remove_3.q.out index a14a8b4..a955da0 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_3.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_3.q.out @@ -69,6 +69,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -149,6 +150,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_4.q.out b/ql/src/test/results/clientpositive/spark/union_remove_4.q.out index 7d94d1d..b0e1f44 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_4.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_4.q.out @@ -62,12 +62,13 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -152,10 +153,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -172,6 +173,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -186,7 +190,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -201,7 +205,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_5.q.out b/ql/src/test/results/clientpositive/spark/union_remove_5.q.out index b31b3e6..edfdd89 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_5.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_5.q.out @@ -70,12 +70,13 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -156,10 +157,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -176,6 +177,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -190,7 +194,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -205,7 +209,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/spark/union_remove_6.q.out b/ql/src/test/results/clientpositive/spark/union_remove_6.q.out index cd36189..2ab5042 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_6.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_6.q.out @@ -65,7 +65,9 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 Stage-1 depends on stages: Stage-2 + Stage-4 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 @@ -176,6 +178,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-3 + Stats-Aggr Operator + Stage: Stage-1 Move Operator tables: @@ -186,6 +191,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2 + Stage: Stage-4 + Stats-Aggr Operator + PREHOOK: query: FROM ( SELECT key, count(1) as `values` from inputTbl1 group by key UNION ALL diff --git a/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out b/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out index 988973f..343afb9 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out @@ -73,7 +73,9 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 Stage-1 depends on stages: Stage-2 + Stage-4 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 @@ -184,6 +186,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-3 + Stats-Aggr Operator + Stage: Stage-1 Move Operator tables: @@ -194,6 +199,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2 + Stage: Stage-4 + Stats-Aggr Operator + PREHOOK: query: FROM ( select * from( SELECT key, count(1) as `values` from inputTbl1 group by key diff --git a/ql/src/test/results/clientpositive/spark/union_remove_7.q.out b/ql/src/test/results/clientpositive/spark/union_remove_7.q.out index 2a5eccf..8df4a0a 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_7.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_7.q.out @@ -67,6 +67,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -161,6 +162,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_8.q.out b/ql/src/test/results/clientpositive/spark/union_remove_8.q.out index 7aee2af..6d3fb11 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_8.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_8.q.out @@ -73,6 +73,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -163,6 +164,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/spark/union_remove_9.q.out b/ql/src/test/results/clientpositive/spark/union_remove_9.q.out index 87ceca2..8f1990b 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_9.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_9.q.out @@ -74,12 +74,13 @@ select * FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -168,10 +169,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -188,6 +189,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Spark #### A masked pattern was here #### Vertices: @@ -198,7 +202,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Spark #### A masked pattern was here #### Vertices: @@ -209,7 +213,7 @@ STAGE PLANS: merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/stats1.q.out b/ql/src/test/results/clientpositive/stats1.q.out index 6f560d3..8dfce75 100644 --- a/ql/src/test/results/clientpositive/stats1.q.out +++ b/ql/src/test/results/clientpositive/stats1.q.out @@ -233,10 +233,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 3 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 1583 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/stats11.q.out b/ql/src/test/results/clientpositive/stats11.q.out index e51f049..bbc476e 100644 --- a/ql/src/test/results/clientpositive/stats11.q.out +++ b/ql/src/test/results/clientpositive/stats11.q.out @@ -91,8 +91,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 1358 #### A masked pattern was here #### @@ -140,8 +140,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 2750 #### A masked pattern was here #### @@ -189,8 +189,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 3 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 4200 #### A masked pattern was here #### @@ -238,8 +238,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 4 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 5812 #### A masked pattern was here #### @@ -405,10 +405,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -520,7 +520,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -529,6 +529,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -539,7 +541,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false bucket_count 2 bucket_field_name key columns key,value @@ -548,6 +550,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1027,10 +1031,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.srcbucket_mapjoin_part numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/stats12.q.out b/ql/src/test/results/clientpositive/stats12.q.out index d5ca59b..36c782c 100644 --- a/ql/src/test/results/clientpositive/stats12.q.out +++ b/ql/src/test/results/clientpositive/stats12.q.out @@ -78,7 +78,7 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' @@ -124,7 +124,7 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' @@ -331,7 +331,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -373,7 +373,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/stats13.q.out b/ql/src/test/results/clientpositive/stats13.q.out index e34f0b7..c287ec5 100644 --- a/ql/src/test/results/clientpositive/stats13.q.out +++ b/ql/src/test/results/clientpositive/stats13.q.out @@ -79,7 +79,7 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' @@ -239,7 +239,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -281,7 +281,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -323,7 +323,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/stats18.q.out b/ql/src/test/results/clientpositive/stats18.q.out index a061846..64a765d 100644 --- a/ql/src/test/results/clientpositive/stats18.q.out +++ b/ql/src/test/results/clientpositive/stats18.q.out @@ -99,8 +99,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 7170 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/stats2.q.out b/ql/src/test/results/clientpositive/stats2.q.out index 5e305d3..fd0047a 100644 --- a/ql/src/test/results/clientpositive/stats2.q.out +++ b/ql/src/test/results/clientpositive/stats2.q.out @@ -15,6 +15,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -49,6 +50,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.analyze_t1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table analyze_t1 partition (ds, hr) select * from srcpart where ds is not null PREHOOK: type: QUERY PREHOOK: Input: default@srcpart diff --git a/ql/src/test/results/clientpositive/stats3.q.out b/ql/src/test/results/clientpositive/stats3.q.out index dd3a95b..23fdeca 100644 --- a/ql/src/test/results/clientpositive/stats3.q.out +++ b/ql/src/test/results/clientpositive/stats3.q.out @@ -87,10 +87,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 11 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/stats6.q.out b/ql/src/test/results/clientpositive/stats6.q.out index b4435f2..e6f4986 100644 --- a/ql/src/test/results/clientpositive/stats6.q.out +++ b/ql/src/test/results/clientpositive/stats6.q.out @@ -166,7 +166,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -208,7 +208,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/stats_noscan_1.q.out b/ql/src/test/results/clientpositive/stats_noscan_1.q.out index 80c3092..6637881 100644 --- a/ql/src/test/results/clientpositive/stats_noscan_1.q.out +++ b/ql/src/test/results/clientpositive/stats_noscan_1.q.out @@ -187,7 +187,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -229,7 +229,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -468,7 +468,7 @@ Table: analyze_srcpart_partial Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -510,7 +510,7 @@ Table: analyze_srcpart_partial Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/stats_noscan_2.q.out b/ql/src/test/results/clientpositive/stats_noscan_2.q.out index 238727d..9ac9bb5 100644 --- a/ql/src/test/results/clientpositive/stats_noscan_2.q.out +++ b/ql/src/test/results/clientpositive/stats_noscan_2.q.out @@ -52,7 +52,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: EXTERNAL_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false EXTERNAL TRUE numFiles 0 numRows -1 diff --git a/ql/src/test/results/clientpositive/stats_partscan_1_23.q.out b/ql/src/test/results/clientpositive/stats_partscan_1_23.q.out index 5488935..eb85c6f 100644 --- a/ql/src/test/results/clientpositive/stats_partscan_1_23.q.out +++ b/ql/src/test/results/clientpositive/stats_partscan_1_23.q.out @@ -78,7 +78,7 @@ Table: analyze_srcpart_partial_scan Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -195,7 +195,7 @@ Table: analyze_srcpart_partial_scan Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/temp_table_display_colstats_tbllvl.q.out b/ql/src/test/results/clientpositive/temp_table_display_colstats_tbllvl.q.out index cfa88ab..8423d50 100644 --- a/ql/src/test/results/clientpositive/temp_table_display_colstats_tbllvl.q.out +++ b/ql/src/test/results/clientpositive/temp_table_display_colstats_tbllvl.q.out @@ -158,7 +158,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false EXTERNAL TRUE bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite @@ -168,6 +168,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -177,7 +179,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false EXTERNAL TRUE bucket_count -1 columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite @@ -187,6 +189,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 + numRows -1 + rawDataSize -1 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_join_filters.q.out b/ql/src/test/results/clientpositive/tez/auto_join_filters.q.out index 8fde41d..f3df656 100644 --- a/ql/src/test/results/clientpositive/tez/auto_join_filters.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_join_filters.q.out @@ -44,7 +44,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 3080335 -Warning: Shuffle Join MERGEJOIN[14][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[11][tables = [a, b]] in Stage 'Reducer 2' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 @@ -330,7 +330,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@myinput1 #### A masked pattern was here #### 3078400 -Warning: Shuffle Join MERGEJOIN[22][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[19][tables = [a, b]] in Stage 'Reducer 2' is a cross product PREHOOK: query: SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value PREHOOK: type: QUERY PREHOOK: Input: default@myinput1 diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out index a275d27..50663db 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out @@ -190,10 +190,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -276,10 +276,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -324,10 +324,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -512,10 +512,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -560,10 +560,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -631,10 +631,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -818,10 +818,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -866,10 +866,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -937,10 +937,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out index 6ac74ca..feb4770 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out @@ -186,10 +186,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -271,10 +271,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -318,10 +318,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -497,10 +497,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -582,10 +582,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -629,10 +629,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -807,10 +807,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -892,10 +892,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -939,10 +939,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1132,10 +1132,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1220,10 +1220,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1267,10 +1267,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1337,10 +1337,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1384,10 +1384,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out index e90af15..b847b55 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out @@ -253,10 +253,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -336,10 +336,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -434,10 +434,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -482,10 +482,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -547,10 +547,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out index 2cb8416..549750e 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out @@ -186,10 +186,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -234,10 +234,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -305,10 +305,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -494,10 +494,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -542,10 +542,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -613,10 +613,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out index abeceb8..d207274 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out @@ -170,10 +170,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -218,10 +218,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -305,10 +305,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -492,10 +492,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -562,10 +562,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -610,10 +610,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -798,10 +798,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -868,10 +868,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -916,10 +916,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out index 8eb9ce5..40b3e3b 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out @@ -186,10 +186,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -234,10 +234,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -321,10 +321,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -508,10 +508,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -578,10 +578,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -626,10 +626,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -814,10 +814,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -884,10 +884,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -932,10 +932,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out index adcc1fa..9d1e3f0 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out @@ -131,7 +131,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -141,6 +141,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -151,7 +153,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -161,6 +163,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -207,7 +211,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -217,6 +221,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -227,7 +233,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -237,6 +243,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -359,7 +367,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -369,6 +377,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -379,7 +389,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -389,6 +399,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -435,7 +447,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -445,6 +457,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -455,7 +469,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -465,6 +479,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -612,7 +628,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -622,6 +638,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -632,7 +650,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -642,6 +660,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -678,7 +698,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -688,6 +708,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -698,7 +720,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key @@ -708,6 +730,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows -1 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out index 2562cb0..3bc951d 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out @@ -203,10 +203,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -251,10 +251,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -338,10 +338,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -386,10 +386,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -576,10 +576,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -624,10 +624,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -695,10 +695,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -743,10 +743,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -933,10 +933,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -981,10 +981,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1052,10 +1052,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1100,10 +1100,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out index 31b0a97..de28f27 100644 --- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out +++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out @@ -203,10 +203,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -251,10 +251,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -338,10 +338,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -386,10 +386,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -576,10 +576,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -624,10 +624,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -695,10 +695,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -743,10 +743,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -935,10 +935,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -983,10 +983,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1054,10 +1054,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1102,10 +1102,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 - numRows 0 + numRows -1 partition_columns ds partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe diff --git a/ql/src/test/results/clientpositive/tez/explainuser_1.q.out b/ql/src/test/results/clientpositive/tez/explainuser_1.q.out index dadcec1..510eb6a 100644 --- a/ql/src/test/results/clientpositive/tez/explainuser_1.q.out +++ b/ql/src/test/results/clientpositive/tez/explainuser_1.q.out @@ -5086,12 +5086,12 @@ Stage-0 |<-Map 1 [SIMPLE_EDGE] Reduce Output Operator [RS_3] sort order: - Statistics:Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Statistics:Num rows: -1 Data size: 171 Basic stats: PARTIAL Column stats: COMPLETE Select Operator [SEL_1] - Statistics:Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Statistics:Num rows: -1 Data size: 171 Basic stats: PARTIAL Column stats: COMPLETE TableScan [TS_0] alias:tgt_rc_merge_test - Statistics:Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Statistics:Num rows: -1 Data size: 171 Basic stats: PARTIAL Column stats: COMPLETE PREHOOK: query: explain select sum(hash(key)), sum(hash(value)) from tgt_rc_merge_test PREHOOK: type: QUERY POSTHOOK: query: explain select sum(hash(key)), sum(hash(value)) from tgt_rc_merge_test diff --git a/ql/src/test/results/clientpositive/tez/insert_into1.q.out b/ql/src/test/results/clientpositive/tez/insert_into1.q.out index 359470b..69a45e8 100644 --- a/ql/src/test/results/clientpositive/tez/insert_into1.q.out +++ b/ql/src/test/results/clientpositive/tez/insert_into1.q.out @@ -367,6 +367,130 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 10 +PREHOOK: query: explain +insert into table insert_into1 values(1, 'abc') +PREHOOK: type: QUERY +POSTHOOK: query: explain +insert into table insert_into1 values(1, 'abc') +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: values__tmp__table__1 + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(tmp_values_col1) (type: int), tmp_values_col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 6 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into1 + + Stage: Stage-3 + Stats-Aggr Operator + +PREHOOK: query: insert into table insert_into1 values(1, 'abc') +PREHOOK: type: QUERY +PREHOOK: Input: default@values__tmp__table__2 +PREHOOK: Output: default@insert_into1 +POSTHOOK: query: insert into table insert_into1 values(1, 'abc') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@values__tmp__table__2 +POSTHOOK: Output: default@insert_into1 +POSTHOOK: Lineage: insert_into1.key EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ] +POSTHOOK: Lineage: insert_into1.value SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ] +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: insert_into1 + Statistics: Num rows: -1 Data size: 120 Basic stats: PARTIAL Column stats: COMPLETE + Select Operator + Statistics: Num rows: -1 Data size: 120 Basic stats: PARTIAL Column stats: COMPLETE + Group By Operator + aggregations: count() + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +11 PREHOOK: query: DROP TABLE insert_into1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into1 diff --git a/ql/src/test/results/clientpositive/tez/stats_noscan_1.q.out b/ql/src/test/results/clientpositive/tez/stats_noscan_1.q.out index aa49526..efbbecc 100644 --- a/ql/src/test/results/clientpositive/tez/stats_noscan_1.q.out +++ b/ql/src/test/results/clientpositive/tez/stats_noscan_1.q.out @@ -187,7 +187,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -229,7 +229,7 @@ Table: analyze_srcpart Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -468,7 +468,7 @@ Table: analyze_srcpart_partial Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 @@ -510,7 +510,7 @@ Table: analyze_srcpart_partial Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 1 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/truncate_column.q.out b/ql/src/test/results/clientpositive/truncate_column.q.out index 2d3e378..17f4fc7 100644 --- a/ql/src/test/results/clientpositive/truncate_column.q.out +++ b/ql/src/test/results/clientpositive/truncate_column.q.out @@ -105,10 +105,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 150 #### A masked pattern was here #### @@ -181,10 +181,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 75 #### A masked pattern was here #### @@ -247,10 +247,10 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 75 #### A masked pattern was here #### @@ -388,11 +388,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false #### A masked pattern was here #### numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 150 #### A masked pattern was here #### @@ -455,11 +455,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - COLUMN_STATS_ACCURATE true + COLUMN_STATS_ACCURATE false #### A masked pattern was here #### numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 75 #### A masked pattern was here #### @@ -608,8 +608,8 @@ Protect Mode: None Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 1 - numRows 0 - rawDataSize 0 + numRows -1 + rawDataSize -1 totalSize 150 #### A masked pattern was here #### diff --git a/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out b/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out index 5ff6607..49efacb 100644 --- a/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out +++ b/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out @@ -115,10 +115,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.test_tab numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct test_tab { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -222,10 +222,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.test_tab numFiles 2 - numRows 0 + numRows -1 partition_columns part partition_columns.types string - rawDataSize 0 + rawDataSize -1 serialization.ddl struct test_tab { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe diff --git a/ql/src/test/results/clientpositive/union_remove_1.q.out b/ql/src/test/results/clientpositive/union_remove_1.q.out index 5fa8caf..4ac2e1a 100644 --- a/ql/src/test/results/clientpositive/union_remove_1.q.out +++ b/ql/src/test/results/clientpositive/union_remove_1.q.out @@ -62,8 +62,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -115,6 +116,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_10.q.out b/ql/src/test/results/clientpositive/union_remove_10.q.out index 228ff9e..86bf188 100644 --- a/ql/src/test/results/clientpositive/union_remove_10.q.out +++ b/ql/src/test/results/clientpositive/union_remove_10.q.out @@ -78,14 +78,15 @@ select * FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-7, Stage-8 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-8, Stage-9 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-7 is a root stage + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 Stage-8 is a root stage + Stage-9 is a root stage STAGE PLANS: Stage: Stage-1 @@ -107,10 +108,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -127,26 +128,29 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-7 + Stage: Stage-8 Map Reduce Map Operator Tree: TableScan @@ -184,7 +188,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-8 + Stage: Stage-9 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_11.q.out b/ql/src/test/results/clientpositive/union_remove_11.q.out index bf38f0a..9670e40 100644 --- a/ql/src/test/results/clientpositive/union_remove_11.q.out +++ b/ql/src/test/results/clientpositive/union_remove_11.q.out @@ -78,12 +78,13 @@ select * FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -153,10 +154,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -173,20 +174,23 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true diff --git a/ql/src/test/results/clientpositive/union_remove_12.q.out b/ql/src/test/results/clientpositive/union_remove_12.q.out index 3ac1d44..ac917f3 100644 --- a/ql/src/test/results/clientpositive/union_remove_12.q.out +++ b/ql/src/test/results/clientpositive/union_remove_12.q.out @@ -70,14 +70,15 @@ FROM inputTbl1 a join inputTbl1 b on a.key=b.key POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-9 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-10 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-10 is a root stage - Stage-9 depends on stages: Stage-10 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-11 is a root stage + Stage-10 depends on stages: Stage-11 STAGE PLANS: Stage: Stage-1 @@ -99,10 +100,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -119,26 +120,29 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-10 + Stage: Stage-11 Map Reduce Local Work Alias -> Map Local Tables: null-subquery2:c-subquery2:a @@ -157,7 +161,7 @@ STAGE PLANS: 0 key (type: string) 1 key (type: string) - Stage: Stage-9 + Stage: Stage-10 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_13.q.out b/ql/src/test/results/clientpositive/union_remove_13.q.out index b12d7cd..95131f5 100644 --- a/ql/src/test/results/clientpositive/union_remove_13.q.out +++ b/ql/src/test/results/clientpositive/union_remove_13.q.out @@ -70,14 +70,15 @@ FROM inputTbl1 a join inputTbl1 b on a.key=b.key POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-9 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-10 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-10 is a root stage - Stage-9 depends on stages: Stage-10 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-11 is a root stage + Stage-10 depends on stages: Stage-11 STAGE PLANS: Stage: Stage-1 @@ -122,10 +123,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -142,26 +143,29 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-10 + Stage: Stage-11 Map Reduce Local Work Alias -> Map Local Tables: null-subquery2:c-subquery2:a @@ -180,7 +184,7 @@ STAGE PLANS: 0 key (type: string) 1 key (type: string) - Stage: Stage-9 + Stage: Stage-10 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_14.q.out b/ql/src/test/results/clientpositive/union_remove_14.q.out index 94f2ff5..ff048f4 100644 --- a/ql/src/test/results/clientpositive/union_remove_14.q.out +++ b/ql/src/test/results/clientpositive/union_remove_14.q.out @@ -72,14 +72,15 @@ FROM inputTbl1 a join inputTbl1 b on a.key=b.key POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-9 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-10 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-10 is a root stage - Stage-9 depends on stages: Stage-10 + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-11 is a root stage + Stage-10 depends on stages: Stage-11 STAGE PLANS: Stage: Stage-1 @@ -101,10 +102,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -121,26 +122,29 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-10 + Stage: Stage-11 Map Reduce Local Work Alias -> Map Local Tables: null-subquery2:c-subquery2:a @@ -159,7 +163,7 @@ STAGE PLANS: 0 key (type: string) 1 key (type: string) - Stage: Stage-9 + Stage: Stage-10 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_15.q.out b/ql/src/test/results/clientpositive/union_remove_15.q.out index 67bdb1e..6195e90 100644 --- a/ql/src/test/results/clientpositive/union_remove_15.q.out +++ b/ql/src/test/results/clientpositive/union_remove_15.q.out @@ -68,8 +68,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -127,6 +128,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_16.q.out b/ql/src/test/results/clientpositive/union_remove_16.q.out index 2444337..7fddc22 100644 --- a/ql/src/test/results/clientpositive/union_remove_16.q.out +++ b/ql/src/test/results/clientpositive/union_remove_16.q.out @@ -66,13 +66,14 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-7 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-8 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-7 is a root stage + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-8 is a root stage STAGE PLANS: Stage: Stage-1 @@ -117,10 +118,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -139,26 +140,29 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-7 + Stage: Stage-8 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_17.q.out b/ql/src/test/results/clientpositive/union_remove_17.q.out index a1b9474..3b058cd 100644 --- a/ql/src/test/results/clientpositive/union_remove_17.q.out +++ b/ql/src/test/results/clientpositive/union_remove_17.q.out @@ -63,6 +63,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -123,6 +124,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/union_remove_18.q.out b/ql/src/test/results/clientpositive/union_remove_18.q.out index 26e52e0..2b05810 100644 --- a/ql/src/test/results/clientpositive/union_remove_18.q.out +++ b/ql/src/test/results/clientpositive/union_remove_18.q.out @@ -66,8 +66,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -125,6 +126,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_19.q.out b/ql/src/test/results/clientpositive/union_remove_19.q.out index d8314f8..89d2f9d 100644 --- a/ql/src/test/results/clientpositive/union_remove_19.q.out +++ b/ql/src/test/results/clientpositive/union_remove_19.q.out @@ -66,8 +66,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -119,6 +120,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -254,8 +258,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -310,6 +315,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -410,8 +418,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -470,6 +479,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_2.q.out b/ql/src/test/results/clientpositive/union_remove_2.q.out index d3522f2..ca9f705 100644 --- a/ql/src/test/results/clientpositive/union_remove_2.q.out +++ b/ql/src/test/results/clientpositive/union_remove_2.q.out @@ -68,9 +68,10 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2, Stage-3 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3, Stage-4 + Stage-2 depends on stages: Stage-0 Stage-3 is a root stage + Stage-4 is a root stage STAGE PLANS: Stage: Stage-1 @@ -122,6 +123,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -140,7 +144,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-3 + Stage: Stage-4 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_20.q.out b/ql/src/test/results/clientpositive/union_remove_20.q.out index 4503726..e2e636a 100644 --- a/ql/src/test/results/clientpositive/union_remove_20.q.out +++ b/ql/src/test/results/clientpositive/union_remove_20.q.out @@ -64,8 +64,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -121,6 +122,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_21.q.out b/ql/src/test/results/clientpositive/union_remove_21.q.out index 98e0166..0c6e875 100644 --- a/ql/src/test/results/clientpositive/union_remove_21.q.out +++ b/ql/src/test/results/clientpositive/union_remove_21.q.out @@ -64,8 +64,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -114,6 +115,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_22.q.out b/ql/src/test/results/clientpositive/union_remove_22.q.out index 4f086f3..0415406 100644 --- a/ql/src/test/results/clientpositive/union_remove_22.q.out +++ b/ql/src/test/results/clientpositive/union_remove_22.q.out @@ -66,8 +66,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -123,6 +124,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -262,8 +266,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -319,6 +324,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_23.q.out b/ql/src/test/results/clientpositive/union_remove_23.q.out index d0e21ba..2777f2c 100644 --- a/ql/src/test/results/clientpositive/union_remove_23.q.out +++ b/ql/src/test/results/clientpositive/union_remove_23.q.out @@ -67,8 +67,9 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2, Stage-4 - Stage-4 is a root stage + Stage-0 depends on stages: Stage-2, Stage-5 + Stage-3 depends on stages: Stage-0 + Stage-5 is a root stage STAGE PLANS: Stage: Stage-1 @@ -154,7 +155,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-4 + Stage: Stage-3 + Stats-Aggr Operator + + Stage: Stage-5 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_24.q.out b/ql/src/test/results/clientpositive/union_remove_24.q.out index f2f6bba..6b3f32f 100644 --- a/ql/src/test/results/clientpositive/union_remove_24.q.out +++ b/ql/src/test/results/clientpositive/union_remove_24.q.out @@ -60,8 +60,9 @@ SELECT * FROM POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -117,6 +118,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_25.q.out b/ql/src/test/results/clientpositive/union_remove_25.q.out index 3858bdc..0eb1954 100644 --- a/ql/src/test/results/clientpositive/union_remove_25.q.out +++ b/ql/src/test/results/clientpositive/union_remove_25.q.out @@ -78,8 +78,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -133,6 +134,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -216,7 +220,7 @@ Table: outputtbl1 Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 2 numRows -1 rawDataSize -1 @@ -273,8 +277,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -328,6 +333,9 @@ STAGE PLANS: name: default.outputtbl2 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -422,7 +430,7 @@ Table: outputtbl2 Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 2 numRows -1 rawDataSize -1 @@ -457,8 +465,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -513,6 +522,9 @@ STAGE PLANS: name: default.outputtbl3 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -612,7 +624,7 @@ Table: outputtbl3 Protect Mode: None #### A masked pattern was here #### Partition Parameters: - COLUMN_STATS_ACCURATE false + COLUMN_STATS_ACCURATE true numFiles 2 numRows -1 rawDataSize -1 diff --git a/ql/src/test/results/clientpositive/union_remove_3.q.out b/ql/src/test/results/clientpositive/union_remove_3.q.out index 40e554a..8f90125 100644 --- a/ql/src/test/results/clientpositive/union_remove_3.q.out +++ b/ql/src/test/results/clientpositive/union_remove_3.q.out @@ -69,6 +69,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 @@ -148,6 +149,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-2 + Stats-Aggr Operator + PREHOOK: query: insert overwrite table outputTbl1 SELECT * FROM ( diff --git a/ql/src/test/results/clientpositive/union_remove_4.q.out b/ql/src/test/results/clientpositive/union_remove_4.q.out index 10a8e89..710713c 100644 --- a/ql/src/test/results/clientpositive/union_remove_4.q.out +++ b/ql/src/test/results/clientpositive/union_remove_4.q.out @@ -62,13 +62,14 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-7 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-8 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-7 is a root stage + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-8 is a root stage STAGE PLANS: Stage: Stage-1 @@ -109,10 +110,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -129,6 +130,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -140,7 +144,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-4 + Stage: Stage-5 Map Reduce Map Operator Tree: TableScan @@ -152,13 +156,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-7 + Stage: Stage-8 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_5.q.out b/ql/src/test/results/clientpositive/union_remove_5.q.out index b59d6da..a8c725a 100644 --- a/ql/src/test/results/clientpositive/union_remove_5.q.out +++ b/ql/src/test/results/clientpositive/union_remove_5.q.out @@ -70,14 +70,15 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-7, Stage-8 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-8, Stage-9 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-7 is a root stage + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 Stage-8 is a root stage + Stage-9 is a root stage STAGE PLANS: Stage: Stage-1 @@ -118,10 +119,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -138,6 +139,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -149,7 +153,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-4 + Stage: Stage-5 Map Reduce Map Operator Tree: TableScan @@ -161,13 +165,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-7 + Stage: Stage-8 Map Reduce Map Operator Tree: TableScan @@ -186,7 +190,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-8 + Stage: Stage-9 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_6.q.out b/ql/src/test/results/clientpositive/union_remove_6.q.out index 67ba74c..b42ef69 100644 --- a/ql/src/test/results/clientpositive/union_remove_6.q.out +++ b/ql/src/test/results/clientpositive/union_remove_6.q.out @@ -64,10 +64,12 @@ insert overwrite table outputTbl2 select * POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-3 depends on stages: Stage-2, Stage-4 + Stage-3 depends on stages: Stage-2, Stage-6 Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 Stage-1 depends on stages: Stage-3 - Stage-4 is a root stage + Stage-5 depends on stages: Stage-1 + Stage-6 is a root stage STAGE PLANS: Stage: Stage-2 @@ -158,6 +160,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-4 + Stats-Aggr Operator + Stage: Stage-1 Move Operator tables: @@ -168,7 +173,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2 - Stage: Stage-4 + Stage: Stage-5 + Stats-Aggr Operator + + Stage: Stage-6 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_6_subq.q.out b/ql/src/test/results/clientpositive/union_remove_6_subq.q.out index 0d1a436..c553ba8 100644 --- a/ql/src/test/results/clientpositive/union_remove_6_subq.q.out +++ b/ql/src/test/results/clientpositive/union_remove_6_subq.q.out @@ -72,10 +72,12 @@ insert overwrite table outputTbl2 select * POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-3 depends on stages: Stage-2, Stage-4 + Stage-3 depends on stages: Stage-2, Stage-6 Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 Stage-1 depends on stages: Stage-3 - Stage-4 is a root stage + Stage-5 depends on stages: Stage-1 + Stage-6 is a root stage STAGE PLANS: Stage: Stage-2 @@ -166,6 +168,9 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 + Stage: Stage-4 + Stats-Aggr Operator + Stage: Stage-1 Move Operator tables: @@ -176,7 +181,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2 - Stage: Stage-4 + Stage: Stage-5 + Stats-Aggr Operator + + Stage: Stage-6 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_7.q.out b/ql/src/test/results/clientpositive/union_remove_7.q.out index 76c678f..889eade 100644 --- a/ql/src/test/results/clientpositive/union_remove_7.q.out +++ b/ql/src/test/results/clientpositive/union_remove_7.q.out @@ -66,8 +66,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-0 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -119,6 +120,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_8.q.out b/ql/src/test/results/clientpositive/union_remove_8.q.out index fe67448..9cf7334 100644 --- a/ql/src/test/results/clientpositive/union_remove_8.q.out +++ b/ql/src/test/results/clientpositive/union_remove_8.q.out @@ -72,9 +72,10 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2, Stage-3 - Stage-2 is a root stage + Stage-0 depends on stages: Stage-1, Stage-3, Stage-4 + Stage-2 depends on stages: Stage-0 Stage-3 is a root stage + Stage-4 is a root stage STAGE PLANS: Stage: Stage-1 @@ -126,6 +127,9 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -144,7 +148,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-3 + Stage: Stage-4 Map Reduce Map Operator Tree: TableScan diff --git a/ql/src/test/results/clientpositive/union_remove_9.q.out b/ql/src/test/results/clientpositive/union_remove_9.q.out index 684091b..f2ed3fa 100644 --- a/ql/src/test/results/clientpositive/union_remove_9.q.out +++ b/ql/src/test/results/clientpositive/union_remove_9.q.out @@ -74,13 +74,14 @@ select * FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-7 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-7 depends on stages: Stage-1, Stage-8 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-7 is a root stage + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-2 depends on stages: Stage-0 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-8 is a root stage STAGE PLANS: Stage: Stage-1 @@ -121,10 +122,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-6 + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -141,26 +142,29 @@ STAGE PLANS: name: default.outputtbl1 Stage: Stage-2 + Stats-Aggr Operator + + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-7 + Stage: Stage-8 Map Reduce Map Operator Tree: TableScan -- 1.7.12.4 (Apple Git-37)